Local Docker Orchestrator
Orchestrating your pipelines to run in Docker.
Last updated
Was this helpful?
Orchestrating your pipelines to run in Docker.
Last updated
Was this helpful?
The local Docker orchestrator is an flavor that comes built-in with ZenML and runs your pipelines locally using Docker.
You should use the local Docker orchestrator if:
you want the steps of your pipeline to run locally in isolated environments.
you want to debug issues that happen when running your pipeline in Docker containers without waiting and paying for remote infrastructure.
To use the local Docker orchestrator, you only need to have installed and running.
To use the local Docker orchestrator, we can register it and use it in our active stack:
You can now run any ZenML pipeline using the local Docker orchestrator:
For example, if you wanted to specify the CPU count available for the Docker image (note: only configurable for Windows), you could write a simple pipeline like the following:
For additional configuration of the Local Docker orchestrator, you can pass LocalDockerOrchestratorSettings
when defining or running your pipeline. Check out the for a full list of available attributes and for more information on how to specify settings. A full list of what can be passed in via the run_args
can be found .
For more information and a full list of configurable attributes of the local Docker orchestrator, check out the .
Note that if you wish to use this orchestrator to run steps on a GPU, you will need to follow to ensure that it works. It requires adding some extra settings customization and is essential to enable CUDA for the GPU to give its full acceleration.