🎶 Imagenette, gentille imagenette,
Imagenette, je te plumerai. 🎶
(Imagenette theme song thanks to Samuel Finlayson)
NB: The Imagenette and Imagewoof datasets have recently (Dec 6 2019) changed. They now have a 70/30 train/valid split. The old versions (which have a much smaller validation set) are still available with the same URLs, but the URLs below point to the new versions. We've also added the new Image网 dataset (see below for details). The leaderboards below been updated using the new datasets, using a strong. Can you beat it?...
This is the home of Imagenette: a subset of 10 easily classified classes from Imagenet (tench, English springer, cassette player, chain saw, church, French horn, garbage truck, gas pump, golf ball, parachute).
'Imagenette' is pronounced just like 'Imagenet', except with a corny inauthentic French accent. If you've seen Peter Sellars in The Pink Panther, then think something like that. It's important to ham up the accent as much as possible, otherwise people might not be sure whether you're refering to "Imagenette" or "Imagenet". (Note to native French speakers: to avoid confusion, be sure to use a corny inauthentic American accent when saying "Imagenet". Think something like the philosophy restaurant skit from Monty Python's The Meaning of Life.)
Here are the datasets: Full size; 320 px; 160 px. The '320 px' and '160 px' versions have their shortest size resized to that size, with their aspect ratio maintained.
Too easy for you? In that case, you might want to try Imagewoof, a subset of 10 classes from Imagenet that aren't so easy to classify, since they're all dog breeds. Here they are: Full size; 320 px; 160 px. The breeds are: Australian terrier, Border terrier, Samoyed, Beagle, Shih-Tzu, English foxhound, Rhodesian ridgeback, Dingo, Golden retriever, Old English sheepdog. (No we will not enter in to any discussion in to whether a dingo is in fact a dog. Any suggestions to the contrary are un-Australian. Thank you for your cooperation.)
Imagewoof too easy for you too?!? Then get your hands on Image网 (pronounced "Imagewang"; 网 means "net" in Chinese)! Here it is: Full size; 320 px; 160 px. Image网 contains Imagenette and Imagewoof combined, but with some twists that make it into a tricky semi-supervised unbalanced classification problem:
- The validation set is the same as Imagewoof (i.e. 30% of Imagewoof images); there are no Imagenette images in the validation set (they're all in the training set)
- Only 10% of Imagewoof images are in the training set!
- The remaining are in the
unsup
("unsupervised") directory, and you can not use their labels in training! - It's even hard to type and hard to say!
- Note that there's no leaderboard for Image网 yet, so feel free to submit your best results to get it started.
I (Jeremy Howard, that is) mainly made Imagenette because I wanted a small vision dataset I could use to quickly see if my algorithm ideas might have a chance of working. They normally don't, but testing them on Imagenet takes a really long time for me to find that out, especially because I'm interested in algorithms that perform particularly well at the end of training.
But I think this can be a useful dataset for others as well.
If you are already using the fastai library, you can download and access these quickly with commands like:
path = untar_data(URLs.IMAGENETTE_160)
where path
now stores the destination to ImageNette-160.
- Try to create a classifier that's as accurate as possible under various constraints (we'll keep leaderboards below, submit your PR with a link to your repo or gist!), such as:
- Within a certain number of epochs: 5, 20, 40, 160
- Within a certain budget on AWS or GCP (use spot or interruptible instances to save money): $0.05, $0.10, $0.25, $0.50, $1.00, $2.00
- Experiment with other low resource problems like transfer learning from small datasets, using semi-supervised learning to help classify small datasets, etc
- Test the impact of using different sized images, either separately, or together as part of training (i.e. progressive resizing)
- Compare your algorithm on easy vs hard small datasets, which are otherwise very similar (Imagenette vs Imagewoof)
- Ensure that you start from random weights - not from pretrained weights.
- Practice your modeling skills on a dataset that's very similar to Imagenet, but much less expensive to deal with
- Do send me a PR with your other applications for this dataset!
- Because there are only 10 categories, the usual "top 5 accuracy" isn't so interesting. So you should generally report top 1 accuracy when using Imagenette
- The best approaches to 5 epoch training often don't scale well to more epochs
- Data augmentation like mixup tends to only help for 80+ epochs
Generally you'll see +/- 1% differences from run to run since it's quite a small validation set. So please only send in contributions that are higher than the reported accuracy >80% of the time. Here's the rules:
- No inference time tricks, e.g. no: TTA, validation size > train size
- Must start with random weights
- Must be one of the size/#epoch combinations listed in the table
- If you have the resources to do so, try to get an average of 5 runs, to get a stable comparison. Use the "# Runs" column to include this (note that
train_imagenette.py
provides a--runs
flag to make this easy) - In the URL column include a link to a notebook, blog post, gist, or similar which explains what you did to get your result, and includes the code you used (or a link to it), including the exact commit, so that others can reproduce your result.
Size (px) | Epochs | URL | Accuracy | # Runs |
---|---|---|---|---|
128 | 5 | fastai2 train_imagenette.py 2020-01 + MaxBlurPool | 85.10% | 5, mean |
128 | 20 | fastai2 train_imagenette.py 2020-01 + MaxBlurPool | 91.57% | 5, mean |
128 | 80 | fastai2 train_imagenette.py 2020-01 | 93.55% | 1 |
128 | 200 | fastai2 train_imagenette.py 2020-01 | 94.24% | 1 |
192 | 5 | fastai2 train_imagenette.py 2020-01 + MaxBlurPool | 86.76% | 5, mean |
192 | 20 | fastai2 train_imagenette.py 2020-01 + MaxBlurPool | 92.50% | 5, mean |
192 | 80 | fastai2 train_imagenette.py 2020-01 | 94.50% | 1 |
192 | 200 | fastai2 train_imagenette.py 2020-01 | 95.03% | 1 |
256 | 5 | fastai2 train_imagenette.py 2020-01 + MaxBlurPool | 86.85% | 5, mean |
256 | 20 | fastai2 train_imagenette.py 2020-01 + MaxBlurPool | 93.53% | 5, mean |
256 | 80 | fastai2 train_imagenette.py 2020-01 | 94.90% | 1 |
256 | 200 | fastai2 train_imagenette.py 2020-01 | 95.11% | 1 |
Size (px) | Epochs | URL | Accuracy | # Runs |
---|---|---|---|---|
128 | 5 | fastai2 train_imagenette.py 2020-01 + MaxBlurPool | 73.37% | 5, mean |
128 | 20 | fastai2 train_imagenette.py 2020-01 + MaxBlurPool | 85.52% | 5, mean |
128 | 80 | fastai2 train_imagenette.py 2020-01 | 87.20% | 1 |
128 | 200 | fastai2 train_imagenette.py 2020-01 | 87.20% | 1 |
192 | 5 | fastai2 train_imagenette.py 2020-01 + MaxBlurPool | 75.94% | 5, mean |
192 | 20 | fastai2 train_imagenette.py 2020-01 + MaxBlurPool | 87.25% | 5, mean |
192 | 80 | fastai2 train_imagenette.py 2020-01 | 89.21% | 1 |
192 | 200 | fastai2 train_imagenette.py 2020-01 | 89.54% | 1 |
256 | 5 | fastai2 train_imagenette.py 2020-01 + MaxBlurPool | 76.87% | 5, mean |
256 | 20 | fastai2 train_imagenette.py 2020-01 + MaxBlurPool | 88.29% | 5, mean |
256 | 80 | fastai2 train_imagenette.py 2020-01 | 90.48% | 1 |
256 | 200 | fastai2 train_imagenette.py 2020-01 | 90.38% | 1 |
Size (px) | Epochs | URL | Accuracy | # Runs |
---|---|---|---|---|
128 | 5 | NA | NA | 1 |
128 | 20 | NA | NA | 1 |
128 | 80 | NA | NA | 1 |
128 | 200 | NA | NA | 1 |
192 | 5 | NA | NA | 1 |
192 | 20 | NA | NA | 1 |
192 | 80 | NA | NA | 1 |
192 | 200 | NA | NA | 1 |
256 | 5 | NA | NA | 1 |
256 | 20 | NA | NA | 1 |
256 | 80 | NA | NA | 1 |
256 | 200 | NA | NA | 1 |