Ask or search…


Welcome to ZenML!
ZenML is an extensible, open-source MLOps framework for creating portable, production-ready machine learning pipelines. By decoupling infrastructure from code, ZenML enables developers across your organization to collaborate more effectively as they develop to production.
ZenML Scarf
ZenML overview
For MLOps Platform Engineers
For Data Scientists
For ML Engineers
ZenML enables MLOps infrastructure experts to define, deploy, and manage sophisticated production environments that are easy to share with colleagues.
  • ZenML Cloud: ZenML Cloud provides a control plane that allows you to deploy a managed ZenML instance and get access to exciting new features such as CI/CD, Model Control Plane, and RBAC.
  • Self-hosted deployment: ZenML can be deployed on any cloud provider and provides many Terraform-based utility functions to deploy other MLOps tools or even entire MLOps stacks:
    # Deploy ZenML to any cloud
    zenml deploy --provider aws
    # Deploy MLOps tools and infrastructure to any cloud
    zenml orchestrator deploy kfp --flavor kubeflow --provider gcp
    # Deploy entire MLOps stacks at once
    zenml stack deploy gcp-vertexai --provider gcp -o kubeflow ...
  • Standardization: With ZenML, you can standardize MLOps infrastructure and tooling across your organization. Simply register your staging and production environments as ZenML stacks and invite your colleagues to run ML workflows on them.
    # Register MLOps tools and infrastructure
    zenml orchestrator register kfp_orchestrator -f kubeflow
    # Register your production environment
    zenml stack register production --orchestrator kubeflow ...
    # Make it available to your colleagues
    zenml stack share production
  • Registering your environments as ZenML stacks also enables you to browse and explore them in a convenient user interface. Try it out at https://www.zenml.io/live-demo!
ZenML Dashboard Stacks View
  • No Vendor Lock-In: Since infrastructure is decoupled from code, ZenML gives you the freedom to switch to a different tooling stack whenever it suits you. By avoiding vendor lock-in, you have the flexibility to transition between cloud providers or services, ensuring that you receive the best performance and pricing available in the market at any time.
    zenml stack set gcp
    python run.py # Run your ML workflows in GCP
    zenml stack set aws
    python run.py # Now your ML workflow runs in AWS
Learn More
Ready to deploy and manage your MLOps infrastructure with ZenML? Here is a collection of pages you can take a look at next:
ZenML gives data scientists the freedom to fully focus on modeling and experimentation while writing code that is production-ready from the get-go.
  • Develop Locally: ZenML allows you to develop ML models in any environment using your favorite tools. This means you can start developing locally, and simply switch to a production environment once you are satisfied with your results.
    python run.py # develop your code locally with all your favorite tools
    zenml stack set production
    python run.py # run on production infrastructure without any code changes
  • Pythonic SDK: ZenML is designed to be as unintrusive as possible. Adding a ZenML @step or @pipeline decorator to your Python functions is enough to turn your existing code into ZenML pipelines:
    from zenml import pipeline, step
    def step_1() -> str:
    return "world"
    def step_2(input_one: str, input_two: str) -> None:
    combined_str = input_one + ' ' + input_two
    def my_pipeline():
    output_step_one = step_1()
    step_2(input_one="hello", input_two=output_step_one)
  • Automatic Metadata Tracking: ZenML automatically tracks the metadata of all your runs and saves all your datasets and models to disk and versions them. Using the ZenML dashboard, you can see detailed visualizations of all your experiments. Try it out at https://www.zenml.io/live-demo!
ZenML Dashboard Details View
ZenML integrates seamlessly with many popular open-source tools, so you can also combine ZenML with other popular experiment tracking tools like Weights & Biases, MLflow, or Neptune for even better reproducibility.
Learn More
Ready to develop production-ready code with ZenML? Here is a collection of pages you can take a look at next:
ZenML empowers ML engineers to take ownership of the entire ML lifecycle end-to-end. Adopting ZenML means fewer handover points and more visibility on what is happening in your organization.
  • ML Lifecycle Management: ZenML's abstractions enable you to manage sophisticated ML setups with ease. After you define your ML workflows as Pipelines and your development, staging, and production infrastructures as Stacks, you can move entire ML workflows to different environments in seconds.
    zenml stack set staging
    python run.py # test your workflows on staging infrastructure
    zenml stack set production
    python run.py # run your workflows in production
  • Reproducibility: ZenML enables you to painlessly reproduce previous results by automatically tracking and versioning all stacks, pipelines, artifacts, and source code. In the ZenML dashboard, you can get an overview of everything that has happened and drill down into detailed lineage visualizations. Try it out at https://www.zenml.io/live-demo!
ZenML Dashboard Overview
ZenML Dashboard Details View
  • Automated Deployments: With ZenML, you no longer need to upload custom Docker images to the cloud whenever you want to deploy a new model to production. Simply define your ML workflow as a ZenML pipeline, let ZenML handle the containerization, and have your model automatically deployed to a highly scalable Kubernetes deployment service like Seldon.
    from zenml.integrations.seldon.steps import seldon_model_deployer_step
    from my_organization.steps import data_loader_step, model_trainer_step
    def my_pipeline():
    data = data_loader_step()
    model = model_trainer_step(data)
Learn More
Ready to manage your ML lifecycles end-to-end with ZenML? Here is a collection of pages you can take a look at next: