/AL-AI-Deeplake

Primary LanguagePythonMozilla Public License 2.0MPL-2.0

We are sunsetting Hub 1.0.

Hub 2.0, an easier-to-use and highly-performant replacement is available in the release/2.0 branch. Please contact us through our Slack Community if you'd like to be granted early access to Hub 2.0.



Docs PyPI version PyPI version CircleCI GitHub issues codecov tweet


Introducing Data 2.0, powered by Hub.
The fastest way to store, access & manage datasets with version-control for PyTorch/TensorFlow. Works locally or on any cloud. Scalable data pipelines.


[ English | Français | 简体中文 | Türkçe | 한글 | Bahasa Indonesia] | Русский]

Note: the translations of this document may not be up-to-date. For the latest version, please check the README in English.

What is Hub for?

Software 2.0 needs Data 2.0, and Hub delivers it. Most of the time Data Scientists/ML researchers work on data management and preprocessing instead of training models. With Hub, we are fixing this. We store your (even petabyte-scale) datasets as single numpy-like array on the cloud, so you can seamlessly access and work with it from any machine. Hub makes any data type (images, text files, audio, or video) stored in cloud usable as fast as if it were stored on premise. With same dataset view, your team can always be in sync.

Hub is being used by Waymo, Red Cross, World Resources Institute, Omdena, and others.

Features

  • Store and retrieve large datasets with version-control
  • Collaborate as in Google Docs: Multiple data scientists working on the same data in sync with no interruptions
  • Access from multiple machines simultaneously
  • Deploy anywhere - locally, on Google Cloud, S3, Azure as well as Activeloop (by default - and for free!)
  • Integrate with your ML tools like Numpy, Dask, Ray, PyTorch, or TensorFlow
  • Create arrays as big as you want. You can store images as big as 100k by 100k!
  • Keep shape of each sample dynamic. This way you can store small and big arrays as 1 array.
  • Visualize any slice of the data in a matter of seconds without redundant manipulations



Visualization of a dataset uploaded to Hub via app.activeloop.ai (free tool).

Getting Started

Work with public or your own data, locally or on any cloud.

Access public data. Fast.

To load a public dataset, one needs to write dozens of lines of code and spend hours accessing and understanding the API as well as downloading the data. With Hub, you only need 2 lines of code, and you can get started working on your dataset in under 3 minutes.

pip3 install hub

To be able to download datasets stores on hub platform, you would need to login: You can register a free account at Activeloop and authenticate locally:

hub register
hub login

# alternatively, add username and password as arguments (use on platforms like Kaggle)
hub login -u username -p password

Access public datasets in Hub by following a straight-forward convention which merely requires a few lines of simple code. Run this excerpt to get the first thousand images in the MNIST database in the numpy array format:

from hub import Dataset

mnist = Dataset("activeloop/mnist")  # loading the MNIST data lazily
# saving time with *compute* to retrieve just the necessary data
mnist["image"][0:1000].compute()

You can find all the other popular datasets on app.activeloop.ai.

Train a model

Load the data and train your model directly. Hub is integrated with PyTorch and TensorFlow and performs conversions between formats in an understandable fashion. Take a look at the example with PyTorch below:

from hub import Dataset
import torch

mnist = Dataset("activeloop/mnist")
# converting MNIST to PyTorch format
mnist = mnist.to_pytorch(lambda x: (x["image"], x["label"]))

train_loader = torch.utils.data.DataLoader(mnist, batch_size=1, num_workers=0)

for image, label in train_loader:
    # Define your training loop here

Create a local dataset

If you want to work on your own data locally, you can start by creating a dataset:

from hub import Dataset, schema
import numpy as np

ds = Dataset(
    "./data/dataset_name",  # file path to the dataset
    shape = (4,),  # follows numpy shape convention
    mode = "w+",  # reading & writing mode
    schema = {  # named blobs of data that may specify types
    # Tensor is a generic structure that can contain any type of data
        "image": schema.Tensor((512, 512), dtype="float"),
        "label": schema.Tensor((512, 512), dtype="float"),
    }
)

# filling the data containers with data (here - zeroes to initialize)
ds["image"][:] = np.zeros((4, 512, 512))
ds["label"][:] = np.zeros((4, 512, 512))
ds.flush()  # executing the creation of the dataset

You can also specify s3://bucket/path, gcs://bucket/path or azure path. Here you can find more information on cloud storage. Also, if you need a publicly available dataset that you cannot find in the Hub, you may file a request. We will enable it for everyone as soon as we can!

Upload your dataset and access it from anywhere in 3 simple steps

  1. Register a free account at Activeloop and authenticate locally:

    activeloop register
    activeloop login
    
    # alternatively, add username and password as arguments (use on platforms like Kaggle)
    activeloop login -u username -p password
  2. Then create a dataset, specifying its name and upload it to your account. For instance:

    from hub import Dataset, schema
    import numpy as np
    
    ds = Dataset(
        "username/dataset_name",
        shape = (4,),
        mode = "w+",
        schema = {
            "image": schema.Tensor((512, 512), dtype="float"),
            "label": schema.Tensor((512, 512), dtype="float"),
        }
    )
    
    ds["image"][:] = np.zeros((4, 512, 512))
    ds["label"][:] = np.zeros((4, 512, 512))
    ds.flush()
  3. Access it from anywhere else in the world, on any device having a command line:

    from hub import Dataset
    
    ds = Dataset("username/dataset_name")

Documentation

For more advanced data pipelines like uploading large datasets or applying many transformations, please refer to our documentation.

Tutorial Notebooks

The examples directory has a series of examples and the notebooks has some notebooks with use cases. Some of the notebooks are listed of below.

Notebook Description
Uploading Images Overview on how to upload and store images on Hub Open In Colab
Uploading Dataframes Overview on how to upload Dataframes on Hub Open In Colab
Uploading Audio Explains how to handle audio data in Hub Open In Colab
Retrieving Remote Data Explains how to retrieve Data Open In Colab
Transforming Data Briefs on how data transformation with Hub Open In Colab
Dynamic Tensors Handling data with variable shape and sizes Open In Colab
NLP using Hub Fine Tuning Bert for CoLA Open In Colab
Getting Started with Text on Hub Overview on using Text datasets in Hub Open In Colab

Use Cases

Why Hub specifically?

There are quite a few dataset management libraries which offer functionality that might seem similar to Hub. In fact, quite a few users migrate data from PyTorch or Tensorflow Datasets to Hub. Here are a few startling differences you will encounter after switching to Hub:

  • the data is provided in chunks, which you may stream from a remote location, instead of downloading all of it at once
  • as only the necessary portion of the dataset is evaluated, you are able to work on the data immediately
  • you are able to store the data that would not fit in your memory in its entirety
  • you may version control and collaborate with multiple users on your datasets across different machines
  • you are equipped with tools that enhance your understanding of the data in a manner of seconds, such as our visualization tool
  • you can easily prepare your data for multiple training libraries at ones (e.g. you can use the same dataset for training with PyTorch and Tensorflow)

Community

Join our Slack community to get help from Activeloop team and other users, as well as stay up-to-date on dataset management/preprocessing best practices.

We'd love your feedback by completing our 3-minute survey.

tweet on Twitter.

As always, thanks to our amazing contributors!

Made with contributors-img.

Please read CONTRIBUTING.md to know how to get started with making contributions to Hub.

Examples

Activeloop's Hub format lets you achieve faster inference at a lower cost. We have 30+ popular datasets already on our platform. These include:

  • COCO
  • CIFAR-10
  • PASCAL VOC
  • Cars196
  • KITTI
  • EuroSAT
  • Caltech-UCSD Birds 200
  • Food101

Check these and many more popular datasets on our visualizer web app and load them directly for model training!

README Badge

Using Hub? Add a README badge to let everyone know:

hub

[![hub](https://img.shields.io/badge/powered%20by-hub%20-ff5a1f.svg)](https://github.com/activeloopai/Hub)

Usage Tracking

By default, we collect anonymous usage data using Bugout (here's the code that does it). It only logs Hub library's own actions and parameters, and no user/ model data is collected.

This helps the Activeloop team to understand how the tool is used and how to deliver maximum value to the community by building features that matter to you. You can easily opt-out of usage tracking during login.

Disclaimers

Similarly to other dataset management packages, Hub is a utility library that downloads and prepares public datasets. We do not host or distribute these datasets, vouch for their quality or fairness, or claim that you have license to use the dataset. It is your responsibility to determine whether you have permission to use the dataset under the dataset's license.

If you're a dataset owner and wish to update any part of it (description, citation, etc.), or do not want your dataset to be included in this library, please get in touch through a GitHub issue. Thanks for your contribution to the ML community!

Acknowledgement

This technology was inspired from our experience at Princeton University and would like to thank William Silversmith @SeungLab with his awesome cloud-volume tool. We are heavy users of Zarr and would like to thank their community for building such a great fundamental block.