Website • Docs • Blog • Twitter • Chat (Community & Support) • Tutorial • Mailing List
Data Version Control or DVC is an open-source tool for data science and machine learning projects. Key features:
- Simple command line Git-like experience. Does not require installing and maintaining any databases. Does not depend on any proprietary online services.
- Management and versioning of datasets and machine learning models. Data is saved in S3, Google cloud, Azure, Alibaba cloud, SSH server, HDFS, or even local HDD RAID.
- Makes projects reproducible and shareable; helping to answer questions about how a model was built.
- Helps manage experiments with Git tags/branches and metrics tracking.
DVC aims to replace spreadsheet and document sharing tools (such as Excel or Google Docs) frequently used as both knowledge repositories and team ledgers. DVC also replaces both ad-hoc scripts to track, move, and deploy different model versions and ad-hoc data file suffixes and prefixes.
Contents
We encourage you to read our Get Started guide to better understand what DVC is and how it can fit your scenarios.
The easiest (but not perfect!) analogy to describe it: DVC is Git (or Git-LFS to be precise) & Makefiles made right and tailored specifically for ML and Data Science scenarios.
Git/Git-LFS
part - DVC helps store and share data artifacts and models, connecting them with a Git repository.Makefile
s part - DVC describes how one data or model artifact was built from other data and code.
DVC usually runs along with Git. Git is used as usual to store and version code (including DVC meta-files). DVC helps to store data and model files seamlessly out of Git, while preserving almost the same user experience as if they were stored in Git itself. To store and share the data cache, DVC supports multiple remotes - any cloud (S3, Azure, Google Cloud, etc) or any on-premise network storage (via SSH, for example).
The DVC pipelines (computational graph) feature connects code and data together. It is possible to explicitly specify all steps required to produce a model: input dependencies including data, commands to run, and output information to be saved. See the quick start section below or the Get Started tutorial to learn more.
Please read Get Started guide for a full version. Common workflow commands include:
Step | Command |
---|---|
Track data | $ git add train.py $ dvc add images.zip |
Connect code and data by commands | $ dvc run -d images.zip -o images/ unzip -q images.zip $ dvc run -d images/ -d train.py -o model.p python train.py |
Make changes and reproduce | $ vi train.py $ dvc repro model.p.dvc |
Share code | $ git add . $ git commit -m 'The baseline model' $ git push |
Share data and ML models | $ dvc remote add myremote -d s3://mybucket/image_cnn $ dvc push |
There are four options to install DVC: pip
, Homebrew, Conda (Anaconda) or an OS-specific package.
Full instructions are available here.
snap install dvc --classic
This corresponds to the latest tagged release.
Add --beta
for the latest tagged release candidate,
or --edge
for the latest master
version.
choco install dvc
brew install dvc
conda install -c conda-forge mamba # installs much faster than conda
mamba install -c conda-forge dvc
Depending on the remote storage type you plan to use to keep and share your data, you might need to install optional dependencies: dvc-s3, dvc-azure, dvc-gdrive, dvc-gs, dvc-oss, dvc-ssh.
pip install dvc
Depending on the remote storage type you plan to use to keep and share your data, you might need to specify
one of the optional dependencies: s3
, gs
, azure
, oss
, ssh
. Or all
to include them all.
The command should look like this: pip install dvc[s3]
(in this case AWS S3 dependencies such as boto3
will be installed automatically).
To install the development version, run:
pip install git+git://github.com/iterative/dvc
Self-contained packages for Linux, Windows, and Mac are available. The latest version of the packages can be found on the GitHub releases page.
sudo wget https://dvc.org/deb/dvc.list -O /etc/apt/sources.list.d/dvc.list
sudo apt-get update
sudo apt-get install dvc
sudo wget https://dvc.org/rpm/dvc.repo -O /etc/yum.repos.d/dvc.repo
sudo yum update
sudo yum install dvc
- Data Engineering tools such as AirFlow, Luigi and others - in DVC data, model and ML pipelines represent a single ML project and focuses on data scientist' experience while data engineering pipeline orchestrates multiple data projects and focuses on efficient execution. DVC project can be used from data pipeline as a single execution step. AirFlow DVC is an example of such integration.
- Git-annex - DVC uses the idea of storing the content of large files (which should not be in a Git repository) in a local key-value store, and uses file hardlinks/symlinks instead of copying/duplicating files.
- Git-LFS - DVC is compatible with any remote storage (S3, Google Cloud, Azure, SSH, etc). DVC also uses reflinks or hardlinks to avoid copy operations on checkouts; thus handling large data files much more efficiently.
- Makefile (and analogues including ad-hoc scripts) - DVC tracks dependencies (in a directed acyclic graph).
- Workflow Management Systems - DVC is a workflow management system designed specifically to manage machine learning experiments. DVC is built on top of Git.
- DAGsHub - online service of Git+DVC repositories with pipeline and metrics visualization and DVC-specific cloud storage.
- DVC Studio - online service of DVC repositories visualisation from DVC team. Also, integrated with CML (CI/CD for ML) for training models in clouds and Kubernetes.
Contributions are welcome! Please see our Contributing Guide for more details.
Want to stay up to date? Want to help improve DVC by participating in our occasional polls? Subscribe to our mailing list. No spam, really low traffic.
This project is distributed under the Apache license version 2.0 (see the LICENSE file in the project root).
By submitting a pull request to this project, you agree to license your contribution under the Apache license version 2.0 to this project.
Iterative, DVC: Data Version Control - Git for Data & Models (2020) DOI:10.5281/zenodo.012345.
Barrak, A., Eghan, E.E. and Adams, B. On the Co-evolution of ML Pipelines and Source Code - Empirical Study of DVC Projects , in Proceedings of the 28th IEEE International Conference on Software Analysis, Evolution, and Reengineering, SANER 2021. Hawaii, USA.