A copy of awesome-papers-fewshot
- Image Classification
- More Direction
- [arXiv 2019] (paper) Generalizing from a Few Examples A Survey on Few-Shot Learning
- [ICLR 2019] (paper code) A Closer Look At Few-shot Classification
One line descriptions: Generate parameters for the classifier or finetune part of the models
-
[ICLR 2017 Meta-learner LSTM Ravi] (paper code) Optimization as a Model for Few-shot Learning
- Use LSTM to generate classifier's parameters
-
[arXiv 2018 REPTILE] (paper) On First-Order Meta-Learning Algorithms
-
[ICLR 2018 SNAIL] (paper) A Simple Neural Attentive Meta- Learner
- Improve the Meta-Learner LSTM, by adding temporal convolution and caual attention to the network.
-
[CVPR 2018] (paper code) Dynamic Few-Shot Visual Learning without Forgetting
-
[CVPR 2018] (paper) Low-Shot Learning with Imprinted Weights
- Passing, generate weights for classifier. (I DONOT like it, the exp only compare to matching networks and "Generative + classifier")
-
[CVPR 2019] (RECOMMENDED!) (paper) Learning to Learn Image Classifiers with Visual Analogy
-
[CVPR 2019] (paper) Dense Classification and Implanting for Few-Shot Learning
-
[ICML 2019] (paper code) LGM-Net: Learning to Generate Matching Networks for Few shot Learning
-
[CVPR 2019] (paper code) TAFE-Net- Task-Aware Feature Embeddings for Low Shot Learning
- Use a meta-learner to generate parameters for the feature extractor
-
[NIPS 2019] (paper code) Incremental Few-Shot Learning with Attention Attractor Networks
- Using normal way to pretrain the backbone on the base classes, then using the base class weights to fintune the classifier on the few-shot episodic network.
- Achieve the normal
-
[ICLR 2019 LEO Vinyals] (RECOMMENDED!) (paper code) Meta-learning with latent embedding optimization
- High dimensional problem is hard to solve in the low-data circumstances, so this work try to bypass the limitations by learning a data-dependent latent low-dimensional latent space of model parameters.
-
[CVPR 2019] (paper code) Generating Classification Weights with GNN Denoising Autoencoders for Few-Shot Learning
- Little better than LEO
-
[NIPS 2018] Delta-encoder: an effective sample synthesis method for few-shot object recognition
-
[ICLR 2019] Meta-learning with differentiable closed-form solvers
- Teach to use tradional machining learning methods
- Most likely no good than LEO
One line descriptions: Generate features to expasion small datasets to large datasets, then fintune.
- [NIPS 2018 Bengio] (paper) MetaGAN An Adversarial Approach to Few-Shot Learning
- [ICCV 2017] (paper code) Low-shot Visual Recognition by Shrinking and Hallucinating Features
- [CVPR 2018] (paper) Low-Shot Learning from Imaginary Data
- [NIPS 2018] (paper) Low-shot Learning via Covariance-Preserving Adversarial Augmentation Networks
- [CVPR 2019] (paper code) Generalized Zero- and Few-Shot Learning via Aligned Variational Autoencoders
- Aiming at cross modal situations
- Using encode image x and class descriptor y, then pull the mean and variance of them together by a loss.
- [CVPR 2019 IDeMe-Net] (paper code) Image Deformation Meta-Networks for One-Shot Learning
- This paper assumes that deformed images may not be visually realistic, they still maintain critical semantic information.
- Pretty good at one-shot on mini-imagenet(59.14%)
One line descriptions: Compute the class representation, then use metric functions to measure the similarity between query sample and each class representaions.
-
[ICML 2012] One-Shot Learning with a Hierarchical Nonparametric Bayesian Model
- Using Hierarchical Bayesian Model after extracted features. Which is similar to build the category graph method in IJCAI 2019.
-
[ICML 2015] Siamese Neural Networks for One-Shot Image Recognition
-
[NIPS 2016] Matching Networks for One Shot Learning
- This methods cast the problem of one-shot learning within the set-to-set framework
-
[NIPS 2017] (RECOMMENDED!) Prototypical Networks for Few-shot Learning
-
[CVPR 2018] Learning to Compare:Relation Network for Few-Shot Learning
- Change metric functions to CNNs
- Provide a clean framework that elegantly encompasses both few and zero-shot learning.
-
[ICLR 2018 Ravi] (paper code) Meta-Learning for Semi-Supervised Few-Shot Classification
- Using soft K-means to refine the prototypes, then using varient ways(training methods) to eliminate the outline points.
- Create new datasets - tiredImagenet
-
[ICLR 2018] Few-Shot Learning with Graph Neural Networks
-
[CVPR 2018] (RECOMMENDED!) Low-Shot Learning With Large-Scale Diffusion
-
[ICLR 2019] (paper) Learning to Propagate Labels-transductive Propagation Network for Few-shot Learning
-
[CVPR 2019] (paper) Edge-Labeling Graph Neural Network for Few-shot Learning
-
[NIPS 2017] Few-Shot Learning Through an Information Retrieval Lens
-
[NIPS 2018] (RECOMMENDED!) TADAM-Task dependent adaptive metric for improved few-shot learning
- In every task, use task representations to finetune the output of each Conv Blocks, like BN functionally.
-
[ECCV 2018] (paper) Dynamic Conditional Networks for FewShot Learning
- Basically same as TADAM(using task representation to finetune the backbones), it use a conditional feature to influence the output of conv layers (Linear combination).
-
[CVPR 2019] (paper code) Finding Task-Relevant Features for Few-Shot Learning by Category Traversal
-
[ICML 2019] (RECOMMENDED!) (paper) TapNet: Neural Network Augmented with Task-Adaptive Projection for Few-Shot Learning
-
[ICML 2019] (paper) (RECOMMENDED!) Infinite Mixture Prototypes for Few-shot Learning
- Point out that data distribution for one class are not uni-model (Verify in my experiments too).
- (Clustering methods) Semi-Supervised methods for prototypical networks. Show this methods even suit for unsupervised situations(protentially).
- Improve on Alphabets dataset, remain or improve on omniglot and mini-imagenet.
-
[ICCV 2019] (paper) Few-Shot Learning with Global Class Representations
- Synthesis new samples to elleviate the data imbalance problem between Base and Novel Classes.
- During training, compute two losses, one is the original losses, the other is the score for the whole classes including noval classes.
-
[IJCAI 2019] (paper code) Prototype Propagation Networks (PPN) for Weakly-supervised Few-shot Learning on Category Graph
- Maually build an category graph, then add parents label's class represention into the child class representations.
-
[CVPR 2019] (RECOMMENDED) (paper code) Variational Prototyping-Encoder- One-Shot Learning with Prototypical Images
- Use encoder to translate the real images to abstract prototypes, such as painted traffic signs, then compare query and sample in the prototypes latent space.
-
[NIPS 2019] (paper) Adaptive Cross-Modal Few-shot Learning
- Using texture information to enhance the performance, which reach a comparable result on mini-imagenet
- Perform well on 1-shot rather than 5-shot or 10-shot
-
[CVPR 2019] (paper) Spot and Learn A Maximum-Entropy Patch Sampler for Few-Shot Image Classification
- Sample parts of the image to from the batch to represent the class.
- One-shot not pretty good(51%)
-
[CVPR 2019] (paper) Baby steps towards few-shot learning with multiple semantics
- Show 4.5 years old baby perform 70% on 1-shot case, adult achieve 99%.
- Add multi-semantic into the task.
- However on 5-shot case LEO perform exceed both this paper and the paper above with no semantics information.
- For 1-shot case, this method achieve 67.2% +- 0.4% compare to 70% of human baby performance.
-
[CVPR 2019] (paper) Few-Shot Learning with Localization in Realistic Settings
- Locate the object in the images first, then classify them.
- Classify in real-world images, somehow not interesting.
-
[NIPS 2019] (paper) Cross Attention Network for Few-shot Classification
- Learn a attention(mask) to pay more attention on the part of the images
- Add transductive inference part
- Pretty good result on mini-imagenet 80.64 +- 0.35% under ResNet-12 (16 conv layers)
-
[CVPR 2019] (paper) Revisiting Local Descriptor based Image-to-Class Measure for Few-shot Learning
- Calculating the similarity between query and class represent feature in feature level, rather than instance level. It seperate original feature in m part and then compute the similarity to the K-nearst class partial features.
- Good Result on mini-ImageNet 71.02 ± 0.64% with Conv4_64F.
-
[CVPR 2019] Large-Scale Few-Shot Learning- Knowledge Transfer With Class Hierarchy
- Aiming at learning large-scale problem, not just on 5 novel class.
- Using the Class Names embeddings(text embedding) to form a class hierarchy.
- Get a pretter higher result than existing methods.
-
[CVPR 2019] (paper) Meta-Transfer Learning for Few-Shot Learning
- Not like it, for the results are not significant, nearly no improve on 5 way 5 shot on mini-ImageNet.
-
[CVPR 2018] (paper) Temporal Hallucinating for Action Recognition with Few Still Images
- Attempt to recall cues from relevant action videos.
- Maybe good at one-shot, not worse than the baseline in 5-shot and 10-shot scenarios.
-
[NIPS 2019] Learning to Propagate for Graph Meta-Learning
- Learns to propagate messages between prototypes of different classes on the graph, so that learning the prototype of each class benefits from the data of other related classes.
- Attention mechanic.
-
[ICCV 2019] Transductive Episodic-Wise Adaptive Metric for Few-Shot Learning
-
[ICCV 2019] (paper Collect and Select: Semantic Alignment Metric Learning for Few-Shot Learning
- Use attention to pick(Select) most relevant part to compare
-
[Nature 子刊 MI 2018] (paper) Continuous Learning of Context-dependent Processing in Neural Networks
- During training a network consecutively for different tasks, OWNs weights are only allowed to be modified in the direction orthogonal to the subspace spanned by all inputs on which the network has been trained (termed input space hereafter). This ensures that new learning processes will not interfere with the learned tasks
-
[ICCV 2019] (RECOMMANDED!) (paper) Diversity with Cooperation: Ensemble Methods for Few-Shot Classification
- New way to solve few-shot learning problems without meta-learing.
-
[ICML 2016] Meta-Learning with Memory-Augmented Neural Networks
This work lead NTM into the image classification, technically, this work should not belong to the few-shot problems. This method can identify the image labels, even the true label of current image are inputed along with the next image.
-
[CVPR 2018] (paper) Memory Matching Networks for One-Shot Image Recognition
-
[ICLR 2019] (paper code) Adaptive Posterior Learning-Few-Shot Learning with a Surprise-Based Memory Module
-
[ICML 2017] Model-Agnostic Meta-Learning for Fast Adaptation of Deep Networks
-
[CVPR 2019] (paper) Task-Agnostic Meta-Learning for Few-shot Learning
A training method force model to learn a unbiased initial model without over-performing on some particular tasks.
- [ICCV 2019] (paper) (RECOMMENDED!) TASK2VEC- Task Embedding for Meta-Learning
- Use Fisher information matrix to judge which backbone is suitable for current task.
- [CVPR 2019 oral] (paper) LaSO-Label-Set Operations networks for multi-label few-shot learning
-
[ICCV 2019] (paper code) Learning Compositional Representations for Few-Shot Recognition
Add additional annotations to the classes.
-
[CVPR 2019] (paper) Few-shot Learning via Saliency-guided Hallucination of Samples
Form segmentations and mix up, aiming at eliminates the back ground noise.
-
[ICCV 2019] (paper) Boosting Few-Shot Visual Learning with Self-Supervision
Self-supervision means to rotate itself, and compute two losses.
-
[NIPS 2019] (paper) Learning to Self-Train for Semi-Supervised Few-Shot Classification
Label the query set for the first run, then retrain the model with the pesudo label for the second run. (Simple but effective)
Basically, we use Omniglot, mini-Imagenet, tiredImagenet, CUB 2011 and full Imagenet for the datasets. We list the latest methods' performs in mini-Imagenet. Welcome contributes to expand the tables of results.
基本上在小样本图像分类领域,主流的数据集为 Omniglot,mini-Imagenet,tired-Imagenet,CUB 和完整的 ImageNet。在这里我们总结了当前已有的方法在 mini-ImageNet 上的表现。 非常欢迎大家来补充呀。
Years | Methods | Backbone | 5-way 1-shot | 5-way 5-shot |
---|---|---|---|---|
2016 | Matching Network | Conv4 | 43.56 +- 0.84% | 55.31% +- 0.73% |
2017 | MAML | Conv4 | 48.7% +- 1.84% | 63.15% +- 0.91% |
2017 | Prototypical Network | Conv4 | 49.42% +- 0.78% | 68.20% +- 0.66% |
2018 | Relation Network | Conv4 | 50.44% +- 0.82% | 65.32% +- 0.70% |
2018 | MetaGAN: An Adversarial Approach to Few-Shot Learning | Conv4 | 46.13+-1.78% | 60.71+-0.89% |
2019 | Incremental Few-Shot Learning with Attention Attractor Networks | ResNet-10 | 54.95+-0.30 | 63.04+-0.30 |
2019 | Adaptive Cross-Modal Few-shot Learning | ResNet-12 | 65.30 ±0.49% | 78.10 ± 0.36% |
2019 | Cross Attention Network for Few-shot Classification | Conv4 | 67.19 ± 0.55 | 80.64 ± 0.35 |
2019 | Learning to Self-Train for Semi-Supervised Few-Shot Classification | ResNet-12 | 70.1 ± 1.9 | 78.7 ± 0.8 |
2019 | Revisiting Local Descriptor based Image-to-Class Measure for Few-shot Learning | Conv-64F | 51.24±0.74% | 71.02±0.64% |
2019 | Few-Shot Learning with Localization in Realistic Settings | ResNet-50 | 49.64±.31% | 69.45±.28% |
2019 | Baby steps towards few-shot learning with multiple semantics | ResNet-12 | 67.2 ± 0.4% | 74.8 ± 0.3% |
2019 | Generating Classification Weights with GNN Denoising Autoencoders for Few-Shot Learning | WRN-28-10 | 62.96+-0.15% | 78.85+-0.10% |
2019 | Spot and Learn: A Maximum-Entropy Patch Sampler for Few-Shot Image Classification | Conv4 | 47.18+-0.83% | 66.41+-0.67% |
2019 | Meta-Transfer Learning for Few-Shot Learning | ResNet-12 | 61.2+-1.8% | 75.5+-0.8% |
2019 | Dense Classification and Implanting for Few-Shot Learning | ResNet-12 | 62.53+-0.19% | 78.95+-0.13% |
2019 | Edge-Labeling Graph Neural Network for Few-shot Learning | Conv4 | 66.85% | |
2019 | Finding Task-Relevant Features for Few-Shot Learning by Category Traversal | COnv4 | 41.62% | 58.77% |
2019 | Few-shot Learning via Saliency-guided Hallucination of Samples | ResNet-12 | 65.30 ±0.49% | 78.10 ± 0.36% |
2018 | Memory Matching Networks for One-Shot Image Recognition | Conv4 | 53.37+-0.48% | 66.97+-0.35% |
2018 | Dynamic Few-Shot Visual Learning without Forgetting | ResNet-12 | 55.45+-0.89% | 70.13+-0.68% |
2018 | Few-Shot Learning with Global Class Representations | Conv4 | 53.21+-0.40% | 72.34+-0.32% |
2018 | LGM-Net: Learning to Generate Matching Networks for Few-Shot Learning | Conv4 | 69.13+-0.35% | 72.28+-0.68% |
2018 | TapNet: Neural Network Augmented with Task-Adaptive Projection for Few-Shot Learning | ResNet-12 | 61.65 ± 0.15% | 76.36 ± 0.10% |
2019 | META-LEARNING WITH LATENT EMBEDDING OPTIMIZATION | WRN-28- 10 | 61.76+-0.08% | 77.59+-0.12% |
2019 | LEARNING TO PROPAGATE LABELS: TRANSDUCTIVE PROPAGATION NETWORK FOR FEW-SHOT LEARNING | Conv4 | 55.51% | 69.86% |
2018 | META-LEARNING FOR SEMI-SUPERVISED FEW-SHOT CLASSIFICATION | Conv4 | 50.09+-0.45% | 64.59+-0.28% |
2018 | A SIMPLE NEURAL ATTENTIVE META-LEARNER | Conv4 | 55.71+-0.99% | 68.88+-0.92% |
2017 | OPTIMIZATION AS A MODEL FOR FEW-SHOT LEARNING | Conv4 | 43.44+-0.77% | 60.60+-0.71% |
2019 | Centroid Networks for Few-Shot Clustering and Unsupervised Few-Shot Classification | Conv4 | 62.6+-0.5% | |
2019 | Infinite Mixture Prototypes for Few-Shot Learning | Conv4 | 49.6+-0.8% | 68.1+-0.8% |
- [CVPR 2017] [Paper] [PyTorch] AGA - Attribute-Guided Augmentation
- Using external depth and pose informations
- [CVPR 2019] [Paper] RepMet-Representative-based Metric Learning for Classification and Few-shot Object Detection
- [CVPR 2019] [Paper] Few-shot Adaptive Faster R-CNN
- [CVPR 2019] [Paper] [Pytorch] [Tensorflow] [Keras] Feature Selective Anchor-Free Module for Single-Shot Object Detection
- [ICCV 2019] [Paper] [PyTorch] Few-shot Object Detection via Feature Reweighting
-
[CVPR 2019] CANet- Class-Agnostic Segmentation Networks with Iterative Refinement and Attentive Few-Shot Learning
-
[AAAI 2019] (paper) Attention-based Multi-Context Guiding for Few-Shot Semantic Segmentation
- Utilize the output of the different layers between query branch and support branch to gain more context informations.
-
[arXiv 2019] AMP-Adaptive Masked Proxies for Few-Shot Segmentation
- Not sure result in this area.
- [ICCV 2019] (paper) Few-Shot Unsupervised Image-to-Image Translation
- [ICCV 2019 best] (paper code) SinGAN: Learning a Generative Model from a Single Natural Image
- [CVPR 2018] Multi-Content GAN for Few-Shot Font Style Transfer
- [NIPS 2019] ( paper code )Few-shot Video-to-Video Synthesis
- [NIPS 2017] Few-Shot Adversarial Domain Adaptation
- [ICCV 2019] Bidirectional One-Shot Unsupervised Domain Mapping
- [ICML 2019] Few-Shot Intent Inference via Meta-Inverse Reinforcement Learning
- [ICCV 2019] Deep Meta Learning for Real-Time Target-Aware Visual Tracking
-
[IJCAI 2019] Incremental Few-Shot Learning for Pedestrian Attribute Recognition
-
[AAAI 2018] AffinityNet- Semi-supervised Few-shot Learning for Disease Type Prediction
- Use few-shot method to enhance oringal disease type prediction
-
[arXiv 2019] (paper) Few-Shot Text Classification with Induction Network
- Introduce dynamic routing to generate better class representations. One real industrial project.
-
[NIPS 2018] Neural Voice Cloning with a Few Samples
-
[IJCAI 2019] Meta-Learning for Low-resource Natural Language Generation in Task-oriented Dialogue Systems
-
[ICCV 2019] (paper) Few-Shot Generalization for Single-Image 3D Reconstruction via Priors
-
[ICCV 2019] (paper) ACMM: Aligned Cross-Modal Memory for Few-Shot Image and Sentence Matching
-
[ICCV 2019] (RECOMMANDED!) Task-Driven Modular Networks for Zero-Shot Compositional Learning
- An interesting usage of a bunch of MLPs.
-
[CVPR 2019] (paper code) SAR Image Classification Using Few-shot Cross-domain Transfer Learning