Polyaxon supports distributed jobs for model training or data processing via several Kubernetes operators.
By default, Polyaxon does not deploy the operators required for running distributed jobs to keep the deployment process lightweight. In order to use a distributed jobs operator, you need to make sure that your namespace/cluster has the operator deployed or you should deploy the operator(s) before starting an execution.
When you start a distributed job, Polyaxon will stream and archive logs from all replicas.
Training operator
In order to run TFJobs/MpiJobs/PytorchJobs/MXNetJobs/PytorchJobs you will need to deploy training operator before executing a job from Polyaxon.
TFJob
Please check the TFJob guide to learn about all details for running TfJobs in Polyaxon.
MpiJob
Please check the MpiJob guide to learn about all details for running MpiJobs in Polyaxon.
PytorchJob
Please check the PytorchJob guide to learn about all details for running PytorchJobs in Polyaxon.
MXNetJob
Please check the MXNetJob guide to learn about all details for running MXNetJobs in Polyaxon.
XGBoostJob
Please check the XGBoostJob guide to learn about all details for running XGBoostJobs in Polyaxon.
Ray
In order to run RayJobs you will need to deploy the kuberay before executing a job from Polyaxon.
Please check the RayJob guide to learn about all details for running RayJobs in Polyaxon.
Dask
Please check the DaskJob guide to learn about all details for running DaskJobs in Polyaxon.