Skip to content

Share and run your AI projects anywhere

Ease handoffs between AI/ML and Application teams

HoW does it WoRK?

Kit is a free open source MLOps tool that packages and versions your AI model, datasets, code, and configuration into a ModelKit that data scientists and developers can use with their preferred tools.

Key feAtuRes

Visit our GitHub repo for a list of all features and our roadmap.

  1. 🎁 Standards-based package

    A ModelKit package includes models, datasets, configurations, and code. Add as much or as little as your project needs.

  2. 🔒 Tamper-proof

    Each ModelKit package includes a SHA digest for itself, and every artifact it holds.

  3. 🏭 Tags and versions

    Each ModelKit is tagged so everyone knows which dataset and model work together.

  4. 🤗 Use with LLM, ML, or AI projects

    ModelKits can be used with any AI, ML, or LLM project - even multi-modal models.

  5. 🤖 Automation for CI/CD

    Pack or unpack a ModelKit locally or as part of your CI/CD workflow for testing, integration, or deployment.

  6. 🏃‍♂️‍ Local dev mode

    Kit's Dev Mode lets your run an LLM locally, configure it, and prompt/chat with it instantly

Why Kit?

Model handoffs are hard.

Moving a model from a Jupyter notebook to an ML tool or development server, then to a production server like Kubernetes is difficult because each tool uses its own packaging mechanism, and requires engineers to repackage the model multiple times. This slows down development and introduces risk.

Kit is an open source MLOps project built to standardize packaging, reproduction, deployment, and tracking of AI / ML models, so it can be run anywhere, just like application code.

Kit solves two big problems:


By building ModelKits on industry standards, anyone (not just data scientists) can participate in the model development lifecycle whether they’re integrating models with their application, experimenting with them locally, or deploying them to production.

ModelKits and Kitfiles work with the tools your team is already using, so you can use the same deployment pipelines and endpoints you’ve hardened with your application development process.

Model traceability and reproducibility–

Unlike Dockerfiles, Kitfiles are a modular package - pull just a part of the ModelKit, like the model or dataset, or pull the whole package with one simple command.

Storing ModelKits in your organization’s registry provides a history of meaningful state changes for auditing. ModelKits are immutable so are perfect for a secure bill-of-materials (SBOM) initiative.


The ModelKit is an OCI compliant package that contains everything needed to integrate with a model, or deploy it to production.

The ModelKit holds the serialized model, dataset, hyperparameters, input / output structure, and validation criteria. Kitfiles define a ModelKit in a modular and easy-to-understand way.

Kit cli

The Kit CLI is a command line interface (CLI) that performs actions on ModelKits.

You can: build and version ModelKits; push or pull them from a model registry; run them locally with a RESTful API we generate for your model automatically, and deploy them to staging or production.

WhAt’s suppoRted?

Kit was designed to work with the tools your team already uses.



Are ModelKits a versioning solution or a packaging solution?

ModelKits do both. With a ModelKit, you can package all the parts of your AI project in one shareable asset, and tag them with a version. ModelKits were designed for the model development lifecycle, where projects are handed off from data science teams to application teams to deployment teams. Versioning and packaging makes it easy for team members to find the datasets and configurations that map to a specific model version. You can read more details about Kit in our overview.

How do I get started with Kit?

The easiest way to get started is to follow our Quick Start, where you’ll learn how to:

  • Package up a model, notebook, and datasets into a single ModelKit you can use with your existing tools
  • Share the ModelKit with others through your public or private registry
  • Grab only the assets you need from the ModelKit for testing, integration, local running, or deployment
Can I see if something changed between ModelKits?

Yes [choir sings hallelujah], each ModelKit includes SHA digests for the ModelKit and every artifact it holds so you can quickly see if something changed between ModelKit versions.

What are the benefits of using Kit?

Increased speed: Teams can work faster with a centralized and versioned package for their AI project coordination. ModelKits eliminate hunting for datasets or code, and make it obvious which datasets and configurations are needed for each model. Handoffs can be automated and executed quickly and with confidence.

Reduced risk: ModelKits are self-verifying. Both the ModelKit itself and all the artifacts added to it are tamper-proof. Anyone can quickly and easily verify when something may have changed.

Improved efficiency: Models stored in ModelKits can be run locally for experimentation or application integration, or packaged for deployment with a single command. Any artifact in a ModelKit can be separately pulled saving time and space on local or shared machines. This makes it easy for data scientists, application developers, and DevOps engineers to find and grab the pieces they need to do their job without being overwhelmed with unnecessary files.

What tools are compatible with Kit?

ModelKits store their assets as OCI-compatible artifacts. This makes them compatible with nearly every development and deployment tool and registry in use today.

Where are ModelKits stored?

ModelKits can be stored in any OCI-compliant registry - for example in a container registry like Docker Hub, or your favorite cloud vendor’s container registry, they can even be stored in an artifact repository like Artifactory.

Where are ModelKits stored?

Yes, it is licensed with the Apache 2.0 license and welcomes all users and contributors. If you’re interested in contributing, let us know.

Are ModelKits a replacement for Docker containers?

No, ModelKits complement containers - in fact, Kit can take a ModelKit and generate a container for the model automatically. However, not all models should be deployed inside containers - sometimes it’s more efficient and faster to deploy an init container linked to the model for deployment. Datasets may also not need to be in containers - many datasets are easier to read and manipulate for training and validation when they’re not in a container. Finally, each container is still separate so even if you do want to put everything in its own container it’s not clear to people outside the AI project which datasets go with which models and which configurations.

Why would I use Kit for versioning instead of Git?

Models and datasets in AI projects are often 10s or 100s of GB in size. Git was designed to work with many small files that can be easily diff’ed between versions. Git treats models and datasets stored in LFS (large file storage) as atomic blobs and can’t differentiate between versions of them. This makes it both inefficient and dangerous since it’s easy for someone to tamper with the models and datasets in the LFS without Git knowing. Finally, once you use LFS, a clone is no longer guaranteed to be the same as the original repo, because the repo refers to an LFS server that is independent of the clone and can change independently.

My MLOps tools do versioning, why do I need Kit?

Kit is the only standards-based and open source solution for packaging and versioning AI project assets. Popular MLOps tools use proprietary and often closed formats to lock you into their ecosystem. This makes handoffs between MLOps tool users and non-MLOps tool users (like your application development and DevOps teams) unnecessarily hard. The future of MLOps tools is still being written, and it’s likely that many will be acquired or shut down and the cost of moving projects from one proprietary format to another is high. By using the OCI standard that’s already supported by nearly every tool on the planet, ModelKits give you a future-proofed solution for packaging and versioning that is compatible with both your MLOps tools and development / DevOps tools so everyone can collaborate regardless of the tools they use.

Is enterprise support available for Kit?

Enterprise support for ModelKits and the Kit CLI is available from Jozu.

Get stArted


Download and install Kit CLI.

Install the CLI

Create a simple manifest file called a Kitfile with your model, dataset and code. Then build and push the ModelKit to a registry for sharing.


Pull the ModelKit into your pipeline, or use kit dev to start working with the model.