The Accelerated Dataloader for PyTorch with Google Cloud Storage lets you connect directly to a GCS bucket as a PyTorch dataset, without pre-loading the data to local disk, and with optimizations to make training up to 3X faster when the dataset consists of many small files (e.g., 100 - 500 KB).
The Accelerated Dataloader for PyTorch implements PyTorch’s dataset primitive that can be used to efficiently load training data from GCS. The library currently supports map-style datasets for random data access patterns and iterable-style datasets for streaming data access patterns.
Furthermore, the Accelerated Dataloader for PyTorch provides a checkpointing interface to conveniently save and load checkpoints directly to and from a Google Cloud Storage (GCS) bucket.
- Python 3.8 or greater is installed (Note: Using 3.12+ is not recommended as PyTorch does not support).
pip install gcs-torch-dataflux
Authentication must be provided to use the Accelerated Dataloader for PyTorch via Application Default Credentials through one of the following methods:
- While running this library on a GCE VM, Application Default Credentials will automatically use the VM’s attached service account by default. More details can be found here.
- Application Default Credentials can also be configured manually as described here. The quickest way is to log in directly using the gcloud CLI:
gcloud auth application-default login
Please checkout the demo
directory for a complete set of examples, which includes a simple starter Jupyter Notebook (hosted by Google Colab) and an end-to-end image segmentation training workload walkthrough. Those examples will help you understand how the Accelerated Dataloader for PyTorch works and how you can integrate it into your own workload.
Before getting started, please make sure you have installed the library and configured authentication following the instructions above.
Both map-style and iterable-style datasets for PyTorch can be constructed by specifying the project name, bucket name and an optional prefix.
from dataflux_pytorch import dataflux_mapstyle_dataset
# Please update these fields.
PROJECT_NAME = "<PROJECT_NAME>"
BUCKET_NAME = "<BUCKET_NAME>"
PREFIX = "<PREFIX>"
# Map-style dataset.
map_dataset = dataflux_mapstyle_dataset.DataFluxMapStyleDataset(
project_name=PROJECT_NAME,
bucket_name=BUCKET_NAME,
config=dataflux_mapstyle_dataset.Config(prefix=PREFIX),
)
# Random access to an object.
sample_object = map_dataset.objects[0]
# Learn about the name and the size (in bytes) of the object.
name = sample_object[0]
size = sample_object[1]
# Iterate over the datasets.
for each_object in map_dataset:
# Raw bytes of the object.
print(each_object)
Similarly, for DataFluxIterableDataset
:
from dataflux_pytorch import dataflux_iterable_dataset
# Iterable-style dataset.
iterable_dataset = dataflux_iterable_dataset.DataFluxIterableDataset(
project_name=PROJECT_ID,
bucket_name=BUCKET_NAME,
config=dataflux_iterable_dataset.Config(prefix=PREFIX),
)
for each_object in iterable_dataset:
# Raw bytes of the object.
print(each_object)
The Accelerated Dataloader for PyTorch offers the flexibility to transform the downloaded raw bytes of data into any format of choice. This is particularly useful since the PyTorch DataLoader
works well with Numpy arrays or PyTorch tensors.
# Assume that you have a bucket with image files and you want
# to load them into Numpy arrays for training.
import io
import numpy
from PIL import Image
transform = lambda img_in_bytes : numpy.asarray(Image.open(io.BytesIO(img_in_bytes)))
map_dataset = dataflux_mapstyle_dataset.DataFluxMapStyleDataset(
project_name=PROJECT_NAME,
bucket_name=BUCKET_NAME,
config=dataflux_mapstyle_dataset.Config(prefix=PREFIX),
data_format_fn=transform,
)
for each_object in map_dataset:
# each_object is now a Numpy array.
print(each_object)
Similarly, for DataFluxIterableDataset
:
iterable_dataset = dataflux_iterable_dataset.DataFluxIterableDataset(
project_name=PROJECT_ID,
bucket_name=BUCKET_NAME,
config=dataflux_mapstyle_dataset.Config(prefix=PREFIX),
data_format_fn=transform,
)
for each_object in dataset:
# each_object is now a Numpy array.
print(each_object)
The Accelerated Dataloader for PyTorch supports fast data loading and allows the user to save and load model checkpoints directly to/from a Google Cloud Storage (GCS) bucket.
import torch
import torchvision
from dataflux_pytorch import dataflux_checkpoint
ckpt = dataflux_checkpoint.DatafluxCheckpoint(
project_name=PROJECT_NAME, bucket_name=BUCKET_NAME
)
CKPT_PATH = "checkpoints/epoch0.ckpt"
model = torchvision.models.resnet50()
with ckpt.writer(CKPT_PATH) as writer:
torch.save(model.state_dict(), writer)
with ckpt.reader(CKPT_PATH) as reader:
read_state_dict = torch.load(reader)
model.load_state_dict(read_state_dict)
The Accelerated Dataloader for PyTorch also provides an integration for PyTorch Lightning, featuring a DatafluxLightningCheckpoint, an implementation of PyTorch Lightning's CheckpointIO.
End to end example and the notebook for the PyTorch Lightning integration can be found in the demo/lightning directory.
from lightning import Trainer
from lightning.pytorch.demos import WikiText2, LightningTransformer
from lightning.pytorch.callbacks import ModelCheckpoint
from torch.utils.data import DataLoader
from dataflux_pytorch.lightning import DatafluxLightningCheckpoint
CKPT = "gcs://BUCKET_NAME/checkpoints/ckpt.ckpt"
dataflux_ckpt = DatafluxLightningCheckpoint(project_name=PROJECT_NAME)
dataset = WikiText2()
dataloader = DataLoader(dataset, num_workers=1)
model = LightningTransformer(vocab_size=dataset.vocab_size)
trainer = Trainer(
plugins=[dataflux_ckpt],
min_epochs=4,
max_epochs=5,
max_steps=10,
accelerator="cpu",
)
trainer.fit(model, dataloader)
trainer.save_checkpoint(CKPT)
We tested the Map-style Dataset's early performance using DLIO benchmark simulations with standard mean file-sizes and dataset sizes. A total of 5 training epochs were simulated. For small files (100KB, 500KB), the Accelerated Dataloader for PyTorch can be 2-3x faster than using GCS native APIs.
File size / count | Tool | Training time (s) |
100 KiB / 500000 files | Direct GCS API calls | 1,299 |
Accelerated Dataloader Map-style Dataset | 515 | |
500 KiB / 2.2m files | Direct GCS API calls | 6,499 |
Accelerated Dataloader Map-style Dataset | 2,058 | |
3 MiB / 50000 files | Direct GCS API calls | 399 |
Accelerated Dataloader Map-style Dataset | 277 | |
150 MiB / 5000 files | Direct GCS API calls | 1,396 |
Accelerated Dataloader Map-style Dataset | 1,173 |
Since the DLIO benchmark doesn’t easily support an implementation of a PyTorch iterable dataset, we implemented a simple training loop that has similar IO behaviors as the DLIO benchmark and used that loop to benchmark the Iterable Datasets.
File size / count | Tool | Training time (s) |
100 KiB / 500000 files | Direct GCS API calls | 1,145 |
Accelerated Dataloader Iterable-style Dataset | 611 | |
500 KiB / 2.2m files | Direct GCS API calls | 5,174 |
Accelerated Dataloader Iterable-style Dataset | 2,503 | |
3 MiB / 50000 files | Direct GCS API calls | 413 |
Accelerated Dataloader Iterable-style Dataset | 384 | |
150 MiB / 5000 files | Direct GCS API calls | 1,225 |
Accelerated Dataloader Iterable-style Dataset | 1,143 |
Note: Within each experiment, all training parameters such as batch size and parallelism are consistent. The team is working on publishing a detailed analysis soon.
Checkpoint benchmarks were taken on single node setups with 48vCPU, 192 GB RAM, 512 GB SSD co-located with a storage bucket in the us-west
region. These results were the average of 10 runs. Checkpoints were tested on models ranging in size from ~76MB to ~24GB.
Checkpoint Type | Checkpoint File Size (MB) | Avg Checkpoint Save Time | Write Throughput (MB/s) |
---|---|---|---|
Without Dataflux | 75.6 | 0.81 | 93.33 |
Dataflux | 75.6 | 0.56 | 135.00 |
Without Dataflux | 298 | 2.87 | 103.98 |
Dataflux | 298 | 1.03 | 289.32 |
Without Dataflux | 2,500 | 25.61 | 97.61 |
Dataflux | 2,500 | 6.25 | 400.00 |
Without Dataflux | 24,200 | 757.10 | 31.96 |
Dataflux | 24,200 | 64.50 | 375.19 |
Checkpoint Type | Checkpoint File Size (MB) | Avg Checkpoint Restore Time | Read Throughput (MB/s) |
---|---|---|---|
Without Dataflux | 75.6 | 2.38 | 31.76 |
Dataflux | 75.6 | 0.51 | 148.24 |
Without Dataflux | 298 | 1.69 | 176.33 |
Dataflux | 298 | 1.03 | 289.32 |
Without Dataflux | 2,500 | 186.57 | 13.40 |
Dataflux | 2,500 | 14.77 | 169.26 |
Without Dataflux | 24,200 | 2,093.52 | 11.56 |
Dataflux | 24,200 | 113.14 | 213.89 |
To optimize listing performance, the Accelerated Dataloader for PyTorch library utilizes a “fast listing” algorithm developed to balance the listing workload among parallelized GCS object listing processes. Therefore, the library will cause more listing API calls to be made than a regular sequential listing, which are charged as Class A operations.
To optimize the download performance of small files, the Accelerated Dataloader for PyTorch library utilizes the GCS Compose API to concatenate a set of smaller objects into a new and larger one in the same bucket under a folder named “dataflux-composed-objects”. The new composite objects will be removed at the end of your training loop but in rare cases that they don’t, you can run this command to clean the composed files.
gcloud storage rm --recursive gs://<my-bucket>/dataflux-composed-objects/
You can turn of this behavior by setting the "disable_compose" parameter to true, or by setting the “max_composite_object_size” parameter to 0 when constructing the dataset. Example:
dataset = dataflux_mapstyle_dataset.DataFluxMapStyleDataset(
project_name=PROJECT_NAME,
bucket_name=BUCKET_NAME,
config=dataflux_mapstyle_dataset.Config(
prefix=PREFIX,
max_composite_object_size=0,
disable_compose=True,
),
)
Note that turning off this behavior may cause the training loop to take significantly longer to complete when working with small files. However, composed download will hit QPS and throughput limits at a lower scale than downloading files directly, so you should disable this behavior when running at high multi-node scales where you are able to hit project QPS or throughput limits without composed download.
To avoid storage charges for retaining the temporary composite objects, consider disabling the Soft Delete retention duration on the bucket.
Due to the quick creation and deletion of composite objects, we recommend that only the Standard storage class is applied to your bucket to minimize cost and maximize performance.
Many machine learning efforts opt for a highly distributed training model leveraging tools such as Pytorch Lightning and Ray. These models are compatible with Dataflux, but can often trigger the rate limits of Google Cloud Storage. This typically manifest in a 429 error, or slower than expected speeds while running distributed operations. Details on specific quotas and limits can be found here.
429 errors acompanied with messages indicating This workload is drawing too much egress bandwidth from Google Cloud Storage
or triggered the Cloud Storage Egress Bandwidth Cap
indicate that the data throughput rate of your workload is exceeding the maximum capacity of your Google Cloud Project. To address these issues, you can take the following steps:
-
Check that other workloads executing within your project are not drawing excess bandwidth. Bandwidth usage can be viewed by following steps here.
-
Follow these instructions to apply for a quota increae of up to 1 Tbps.
-
If more than 1 Tpbs is required, contact your Technical Account Manager or Google representative to file a request on your behalf. This request should specify that you wish to increase the GCS bandwidth caps on your project.
-
Adjust the
listing_retry_config
anddownload_retry_config
options to tune your retry backoff and maximize performance.from google.cloud.storage.retry import DEFAULT_RETRY dataset = dataflux_mapstyle_dataset.DataFluxMapStyleDataset( project_name=PROJECT_NAME, bucket_name=BUCKET_NAME, config=dataflux_mapstyle_dataset.Config( prefix=PREFIX, max_composite_object_size=0, list_retry_config=DEFAULT_RETRY.with_deadline(300.0).with_delay( initial=1.0, multiplier=1.2, maximum=45.0 ), download_retry_config=DEFAULT_RETRY.with_deadline(600.0).with_delay( initial=1.0, multiplier=1.5, maximum=90.0 ), ), )
QPS limits can trigger 429 errors with a body message indicating Too many Requests
, but more commonly manifest in slower than expected execution times. QPS bottlenecks are more common when operating on high volumes of small files. Note that bucket QPS limits will naturally scale over time, so allowing a grace period for warmup can often lead to faster performance. To get more detail on the performance of a target bucket, look at the Observability
tab when viewing your bucket from the Cloud Console.
If your workload is failing with messages similar to TooManyRequests: 429 GET https://storage.googleapis.com/download/storage/v1/b/<MY-BUCKET>/o/dataflux-composed-objects%2Fa80727ae-7bc9-4ba3-8f9b-13ff001d6418
that contain the "dataflux-composed-objects" keyword, disabling composed objects is the best first troubleshooting step. This can reduce QPS load brought on by the compose API when used at scale.
We welcome your feedback, issues, and bug fixes. We have a tight roadmap at this time so if you have a major feature or change in functionality you'd like to contribute, please open a GitHub Issue for discussion prior to sending a pull request. Please see CONTRIBUTING for more information on how to report bugs or submit pull requests.
This project has adopted the Google Open Source Code of Conduct. Please see code-of-conduct.md for more information.
The Accelerated Dataloader for PyTorch library has an Apache License 2.0. Please see the LICENSE file for more information.