Augmentor is an image augmentation library in Python for machine learning. It aims to make image augmentation platform and framework independent, more convenient, less error prone, and reproducible. It employs a stochastic approach using building blocks that allow for operations to be pieced together in a pipeline.
Augmentor is written in Python. A Julia version of the package is also being developed as a sister project and is available here.
Install using pip
from the command line:
pip install Augmentor
See the documentation for building from source. To upgrade from a previous version, use pip install Augmentor --upgrade
.
Complete documentation can be found on Read the Docs: http://augmentor.readthedocs.io/
The purpose of Augmentor is to automate image augmentation (artificial data generation) in order to expand datasets as input for machine learning algorithms, especially neural networks and deep learning.
The package works by building an augmentation pipeline where you define a series of operations to perform on a set of images. Operations, such as rotations or transforms, are added one by one to create an augmentation pipeline: when complete, the pipeline can be executed and an augmented dataset is created.
To begin, instantiate a Pipeline
object that points to a directory on your file system:
import Augmentor
p = Augmentor.Pipeline("/path/to/images")
You can then add operations to the Pipeline object p
as follows:
p.rotate(probability=0.7, max_left=10, max_right=10)
p.zoom(probability=0.5, min_scale=1.1, max_scale=1.5)
Every function requires you to specify a probability, which is used to decide if an operation is applied to an image as it is passed through the augmentation pipeline.
Once you have created a pipeline, you can sample from it like so:
p.sample(10000)
which will generate 10,000 augmented images based on your specifications. By default these will be written to the disk in a directory named output
relative to the path specified when initialising the p
pipeline object above.
If you do not wish to save to disk, you can use a generator (in this case with Keras):
g = p.keras_generator(batch_size=128)
images, labels = next(g)
which returns a batch of images of size 128 and their corresponding labels. Generators return data indefinitely, and can be used to train neural networks with augmented data on the fly.
Alternatively, you can integrate it with PyTorch:
import torchvision
transforms = torchvision.transforms.Compose([
p.torch_transform(),
torchvision.transforms.ToTensor(),
])
Using elastic distortions, one image can be used to generate many images that are real-world feasible and label preserving:
Input Image | Augmented Images | |
---|---|---|
→ |
The input image has a 1 pixel black border to emphasise that you are getting distortions without changing the size or aspect ratio of the original image, and without any black/transparent padding around the newly generated images.
The functionality can be more clearly seen here:
Original Image[1] | Random distortions applied |
---|---|
There are a total of 12 different types of perspective transform available. Four of the most common are shown below.
Tilt Left | Tilt Right | Tilt Forward | Tilt Backward |
---|---|---|---|
The remaining eight types of transform are as follows:
Skew Type 0 | Skew Type 1 | Skew Type 2 | Skew Type 3 |
---|---|---|---|
Skew Type 4 | Skew Type 5 | Skew Type 6 | Skew Type 7 |
---|---|---|---|
Rotations by default preserve the file size of the original images:
Original Image | Rotated 10 degrees, automatically cropped |
---|---|
Compared to rotations by other software:
Original Image | Rotated 10 degrees |
---|---|
Shearing will also automatically crop the correct area from the sheared image, so that you have an image with no black space or padding.
Original image | Shear (x-axis) 20 degrees | Shear (y-axis) 20 degrees |
---|---|---|
Compare this to how this is normally done:
Original image | Shear (x-axis) 20 degrees | Shear (y-axis) 20 degrees |
---|---|---|
Cropping can also be handled in a manner more suitable for machine learning image augmentation:
Original image | Random crops + resize operation |
---|---|
Random Erasing is a technique used to make models robust to occlusion. This may be useful for training neural networks used in object detection in navigation scenarios, for example.
Original image[2] | Random Erasing |
---|---|
See http://augmentor.readthedocs.io/en/master/code.html#Augmentor.Pipeline.Pipeline.random_erasing for usage.
With only a few operations, a single image can be augmented to produce large numbers of new, label-preserving samples:
Original image | Distortions + mirroring |
---|---|
In the example above, we have applied three operations: first we randomly distort the image, then we flip it horizontally with a probability of 0.5 and then vertically with a probability of 0.5. We then sample from this pipeline 100 times to create 100 new data.
p.random_distortion(probability=1, grid_width=4, grid_height=4, magnitude=8)
p.flip_left_right(probability=0.5)
p.flip_top_bottom(probability=0.5)
p.sample(100)
Augmentor can be used as a replacement for Keras' augmentation functionality. Augmentor can create a generator which produces augmented data indefinitely, according to the pipeline you have defined. See the following notebooks for details:
- Reading images from a local directory, augmenting them at run-time, and using a generator to pass the augmented stream of images to a Keras convolutional neural network, see
Augmentor_Keras.ipynb
- Augmenting data in-memory (in array format) and using a generator to pass these new images to the Keras neural network, see
Augmentor_Keras_Array_Data.ipynb
Augmentor allows for pipelines to be defined per class. That is, you can define different augmentation strategies on a class-by-class basis for a given classification problem.
See an example of this in the following Jupyter notebook: Per_Class_Augmentation_Strategy.ipynb
Let's perform an augmentation task on a single image, demonstrating the pipeline and several features of Augmentor.
First import the package and initialise a Pipeline object by pointing it to a directory containing your images:
import Augmentor
p = Augmentor.Pipeline("/home/user/augmentor_data_tests")
Now you can begin adding operations to the pipeline object:
p.rotate90(probability=0.5)
p.rotate270(probability=0.5)
p.flip_left_right(probability=0.8)
p.flip_top_bottom(probability=0.3)
p.crop_random(probability=1, percentage_area=0.5)
p.resize(probability=1.0, width=120, height=120)
Once you have added the operations you require, you can sample images from this pipeline:
p.sample(100)
Some sample output:
Input Image[3] | Augmented Images | |
---|---|---|
→ |
The augmented images may be useful for a boundary detection task, for example.
Augmentor is made available under the terms of the MIT Licence. See Licence.md
.
[1] Checkerboard image obtained from Wikimedia Commons and is in the public domain: https://commons.wikimedia.org/wiki/File:Checkerboard_pattern.svg
[2] Street view image is in the public domain: http://stokpic.com/project/italian-city-street-with-shoppers/
[3] Skin lesion image obtained from the ISIC Archive:
- Image id = 5436e3abbae478396759f0cf
- Download: https://isic-archive.com:443/api/v1/image/5436e3abbae478396759f0cf/download
You can use urllib
to obtain the skin lesion image in order to reproduce the augmented images above:
>>> from urllib import urlretrieve
>>> im_url = "https://isic-archive.com:443/api/v1/image/5436e3abbae478396759f0cf/download"
>>> urlretrieve(im_url, "ISIC_0000000.jpg")
('ISIC_0000000.jpg', <httplib.HTTPMessage instance at 0x7f7bd949a950>)
Note: For Python 3, use from urllib.request import urlretrieve
.
If you find this package useful and wish to cite it, you can use
Marcus D. Bloice, Christof Stocker, and Andreas Holzinger, Augmentor: An Image Augmentation Library for Machine Learning, arXiv preprint arXiv:1708.04680, https://arxiv.org/abs/1708.04680, 2017.
Click the preview below to view a video demonstration of Augmentor in use: