Awesome-Forgetting-in-Deep-Learning

Awesome

A comprehensive list of papers about 'A Comprehensive Survey of Forgetting in Deep Learning Beyond Continual Learning'.

Abstract

Forgetting refers to the loss or deterioration of previously acquired information or knowledge. While the existing surveys on forgetting have primarily focused on continual learning, forgetting is a prevalent phenomenon observed in various other research domains within deep learning. Forgetting manifests in research fields such as generative models due to generator shifts, and federated learning due to heterogeneous data distributions across clients. Addressing forgetting encompasses several challenges, including balancing the retention of old task knowledge with fast learning of new tasks, managing task interference with conflicting goals, and preventing privacy leakage, etc. Moreover, most existing surveys on continual learning implicitly assume that forgetting is always harmful. In contrast, our survey argues that forgetting is a double-edged sword and can be beneficial and desirable in certain cases, such as privacy-preserving scenarios. By exploring forgetting in a broader context, we aim to present a more nuanced understanding of this phenomenon and highlight its potential advantages. Through this comprehensive survey, we aspire to uncover potential solutions by drawing upon ideas and approaches from various fields that have dealt with forgetting. By examining forgetting beyond its conventional boundaries, in future work, we hope to encourage the development of novel strategies for mitigating, harnessing, or even embracing forgetting in real applications.

Citation

If you find our paper or this resource helpful, please consider cite:

@article{Forgetting_Survey_2023,
  title={A Comprehensive Survey of Forgetting in Deep Learning Beyond Continual Learning},
  author={Zhenyi Wang and Enneng Yang and Li Shen and Heng Huang},
  journal={arXiv preprint arXiv:2307.09218},
  year={2023}
}

Thanks!


Framework


Harmful Forgetting

Harmful forgetting occurs when we desire the machine learning model to retain previously learned knowledge while adapting to new tasks, domains, or environments. In such cases, it is important to prevent and mitigate knowledge forgetting.

Problem Setting Goal Source of forgetting
Continual Learning learn non-stationary data distribution without forgetting previous knowledge data-distribution shift during training
Foundation Model unsupervised learning on large-scale unlabeled data data-distribution shift in pre-training, fine-tuning
Domain Adaptation adapt to target domain while maintaining performance on source domain target domain sequentially shift over time
Test-time Adaptation mitigate the distribution gap between training and testing adaptation to the test data distribution during testing
Meta-Learning learn adaptable knowledge to new tasks incrementally meta-learn new classes / task-distribution shift
Generative Model learn a generator to appriximate real data distribution generator shift/data-distribution shift
Reinforcement Learning maximize accumulate rewards state, action, reward and state transition dynamics
Federated Learning decentralized training without sharing data model average; non-i.i.d data; data-distribution shift

Links: Forgetting in Continual Learning | Forgetting in Foundation Models | Forgetting in Domain Adaptation | Forgetting in Test-Time Adaptation |
Forgetting in Meta-Learning |
Forgetting in Generative Models | Forgetting in Reinforcement Learning | Forgetting in Federated Learning


Forgetting in Continual Learning

The goal of continual learning (CL) is to learn on a sequence of tasks without forgetting the knowledge on previous tasks.

Links: Task-aware CL | Task-free CL | Online CL | Semi-supervised CL | Few-shot CL | Unsupervised CL | Theoretical Analysis

Task-aware CL

Task-aware CL focuses on addressing scenarios where explicit task definitions, such as task IDs or labels, are available during the CL process. Existing methods on task-aware CL have explored five main branches: Memory-based Methods | Architecture-based Methods | Regularization-based Methods | Subspace-based Methods | Bayesian Methods.

Memory-based Methods

Memory-based method keeps a memory buffer that stores the examples/knowledges from previous tasks and replay those examples during learning new tasks.

Paper Title Year Conference/Journal
Error Sensitivity Modulation based Experience Replay: Mitigating Abrupt Representation Drift in Continual Learning 2023 ICLR
A Model or 603 Exemplars: Towards Memory-Efficient Class-Incremental Learning 2023 ICLR
DualHSIC: HSIC-Bottleneck and Alignment for Continual Learning 2023 ICML
Regularizing Second-Order Influences for Continual Learning 2023 CVPR
Class-Incremental Exemplar Compression for Class-Incremental Learning 2023 CVPR
Class-Incremental Learning using Diffusion Model for Distillation and Replay 2023 Arxiv
On the Effectiveness of Lipschitz-Driven Rehearsal in Continual Learning 2022 NeurIPS
Exploring Example Influence in Continual Learning 2022 NeurIPS
Navigating Memory Construction by Global Pseudo-Task Simulation for Continual Learning 2022 NeurIPS
Learning Fast, Learning Slow: A General Continual Learning Method based on Complementary Learning System 2022 ICLR
Information-theoretic Online Memory Selection for Continual Learning 2022 ICLR
Memory Replay with Data Compression for Continual Learning 2022 ICLR
Improving Task-free Continual Learning by Distributionally Robust Memory Evolution 2022 ICML
GCR: Gradient Coreset based Replay Buffer Selection for Continual Learning 2022 CVPR
RMM: Reinforced Memory Management for Class-Incremental Learning 2021 NeurIPS
Rainbow Memory: Continual Learning with a Memory of Diverse Samples 2021 CVPR
Always Be Dreaming: A New Approach for Data-Free Class-Incremental Learning 2021 ICCV
Using Hindsight to Anchor Past Knowledge in Continual Learning 2021 AAAI
Improved Schemes for Episodic Memory-based Lifelong Learning 2020 NeurIPS
Dark Experience for General Continual Learning: a Strong, Simple Baseline 2020 NeurIPS
La-MAML: Look-ahead Meta Learning for Continual Learning 2020 NeurIPS
Brain-inspired replay for continual learning with artificial neural networks 2020 Nature Communications
LAMOL: LAnguage MOdeling for Lifelong Language Learning 2020 ICLR
Mnemonics Training: Multi-Class Incremental Learning without Forgetting 2020 CVPR
GDumb: A Simple Approach that Questions Our Progress in Continual Learning 2020 ECCV
Continual Learning with Tiny Episodic Memories 2019 ICML
Efficient lifelong learning with A-GEM 2019 ICLR
Learning to Learn without Forgetting by Maximizing Transfer and Minimizing Interference 2019 ICLR
Large Scale Incremental Learning 2019 CVPR
On Tiny Episodic Memories in Continual Learning 2019 Arxiv
Progress & Compress: A scalable framework for continual learning 2018 ICML
Gradient Episodic Memory for Continual Learning 2017 NeurIPS
Continual Learning with Deep Generative Replay 2017 NeurIPS
iCaRL: Incremental Classifier and Representation Learning 2017 CVPR
Catastrophic forgetting, rehearsal and pseudorehearsal Connection Science 1995
Architecture-based Methods

The architecture-based approach avoids forgetting by reducing parameter sharing between tasks or adding parameters to new tasks.

Paper Title Year Conference/Journal
CLR: Channel-wise Lightweight Reprogramming for Continual Learning 2023 ICCV
Parameter-Level Soft-Masking for Continual Learning 2023 ICML
Continual Learning on Dynamic Graphs via Parameter Isolation 2023 SIGIR
Heterogeneous Continual Learning 2023 CVPR
Dense Network Expansion for Class Incremental Learning 2023 CVPR
Achieving a Better Stability-Plasticity Trade-off via Auxiliary Networks in Continual Learning 2023 CVPR
Forget-free Continual Learning with Winning Subnetworks 2022 ICML
NISPA: Neuro-Inspired Stability-Plasticity Adaptation for Continual Learning in Sparse Networks 2022 ICML
Continual Learning with Filter Atom Swapping 2022 ICLR
SparCL: Sparse Continual Learning on the Edge 2022 NeurIPS
Learning Bayesian Sparse Networks with Full Experience Replay for Continual Learning 2022 CVPR
FOSTER: Feature Boosting and Compression for Class-Incremental Learning 2022 ECCV
BNS: Building Network Structures Dynamically for Continual Learning 2021 NeurIPS
DER: Dynamically Expandable Representation for Class Incremental Learning 2021 CVPR
Adaptive Aggregation Networks for Class-Incremental Learning 2021 CVPR
BatchEnsemble: an Alternative Approach to Efficient Ensemble and Lifelong Learning 2020 ICLR
Calibrating CNNs for Lifelong Learning 2020 NeurIPS
Compacting, Picking and Growing for Unforgetting Continual Learning 2019 NeurIPS
Superposition of many models into one 2019 NeurIPS
Reinforced Continual Learning 2018 NeurIPS
Progress & Compress: A scalable framework for continual learning 2018 ICML
Overcoming Catastrophic Forgetting with Hard Attention to the Task 2018 ICML
Lifelong Learning with Dynamically Expandable Networks 2018 ICLR
PackNet: Adding Multiple Tasks to a Single Network by Iterative Pruning 2018 CVPR
Expert Gate: Lifelong Learning with a Network of Experts 2017 CVPR
Progressive Neural Networks 2016 Arxiv
Regularization-based Methods

Regularization-based approaches avoid forgetting by penalizing updates of important parameters or distilling knowledge with previous model as a teacher.

Paper Title Year Conference/Journal
Overcoming Catastrophic Forgetting in Incremental Object Detection via Elastic Response Distillation 2022 CVPR
Natural continual learning: success is a journey, not (just) a destination 2021 NeurIPS
CPR: Classifier-Projection Regularization for Continual Learning 2021 ICLR
Continual Learning with Node-Importance based Adaptive Group Sparse Regularization 2020 NeurIPS
Uncertainty-based Continual Learning with Adaptive Regularization 2019 NeurIPS
Efficient Lifelong Learning with A-GEM 2019 ICLR
Riemannian Walk for Incremental Learning: Understanding Forgetting and Intransigence 2018 ECCV
Memory Aware Synapses: Learning what (not) to forget 2018 ECCV
Overcoming catastrophic forgetting in neural networks 2017 Arxiv
Continual Learning Through Synaptic Intelligence 2017 ICML
Learning without Forgetting 2017 TPAMI
Subspace-based Methods

Subspace-based methods perform CL in multiple disjoint subspaces to avoid interference between multiple tasks.

Paper Title Year Conference/Journal
Building a Subspace of Policies for Scalable Continual Learning 2023 ICLR
Rethinking Gradient Projection Continual Learning: Stability / Plasticity Feature Space Decoupling 2023 CVPR
Continual Learning with Scaled Gradient Projection 2023 AAAI
SketchOGD: Memory-Efficient Continual Learning 2023 Arxiv
Beyond Not-Forgetting: Continual Learning with Backward Knowledge Transfer 2022 NeurIPS
TRGP: Trust Region Gradient Projection for Continual Learning 2022 ICLR
Continual Learning with Recursive Gradient Optimization 2022 ICLR
Balancing Stability and Plasticity through Advanced Null Space in Continual Learning 2022 ECCV
Adaptive Orthogonal Projection for Batch and Online Continual Learning 2022 AAAI
Natural continual learning: success is a journey, not (just) a destination 2021 NeurIPS
Flattening Sharpness for Dynamic Gradient Projection Memory Benefits Continual Learning 2021 NeurIPS
Gradient Projection Memory for Continual Learning 2021 ICLR
Training Networks in Null Space of Feature Covariance for Continual Learning 2021 CVPR
Generalisation Guarantees For Continual Learning With Orthogonal Gradient Descent 2021 Arxiv
Defeating Catastrophic Forgetting via Enhanced Orthogonal Weights Modification 2021 Arxiv
Continual Learning in Low-rank Orthogonal Subspaces 2020 NeurIPS
Orthogonal Gradient Descent for Continual Learning 2020 AISTATS
Generalisation Guarantees for Continual Learning with Orthogonal Gradient Descent 2020 Arxiv
Generative Feature Replay with Orthogonal Weight Modification for Continual Learning 2020 Arxiv
Continual Learning of Context-dependent Processing in Neural Networks 2019 Nature Machine Intelligence
Bayesian Methods

Bayesian methods provide a principled probabilistic framework for addressing Forgetting.

Paper Title Year Conference/Journal
A Probabilistic Framework for Modular Continual Learning 2023 Arxiv
Online Continual Learning on Class Incremental Blurry Task Configuration with Anytime Inference 2022 ICLR
Continual Learning via Sequential Function-Space Variational Inference 2022 ICML
Generalized Variational Continual Learning 2021 ICLR
Variational Auto-Regressive Gaussian Processes for Continual Learning 2021 ICML
Bayesian Structural Adaptation for Continual Learning 2021 ICML
Continual Learning using a Bayesian Nonparametric Dictionary of Weight Factors 2021 AISTATS
Posterior Meta-Replay for Continual Learning 2021 NeurIPS
Natural continual learning: success is a journey, not (just) a destination 2021 NeurIPS
Continual Learning with Adaptive Weights (CLAW) 2020 ICLR
Uncertainty-guided Continual Learning with Bayesian Neural Networks 2020 ICLR
Functional Regularisation for Continual Learning with Gaussian Processes 2020 ICLR
Continual Deep Learning by Functional Regularisation of Memorable Past 2020 NeurIPS
Variational Continual Learning 2018 ICLR
Online Structured Laplace Approximations for Overcoming Catastrophic Forgetting 2018 NeurIPS
Overcoming Catastrophic Forgetting by Incremental Moment Matching 2017 NeurIPS

Task-free CL

Task-free CL refers to a specific scenario that the learning system does not have access to any explicit task information.

Paper Title Year Conference/Journal
Online Bias Correction for Task-Free Continual Learning 2023 ICLR
Task-Free Continual Learning via Online Discrepancy Distance Learning 2022 NeurIPS
Improving Task-free Continual Learning by Distributionally Robust Memory Evolution 2022 ICML
VariGrow: Variational architecture growing for task-agnostic continual learning based on Bayesian novelty 2022 ICML
Gradient-based Editing of Memory Examples for Online Task-free Continual Learning 2021 NeurIPS
Continuous Meta-Learning without Tasks 2020 NeurIPS
A Neural Dirichlet Process Mixture Model for Task-Free Continual Learning 2020 ICLR
Online Continual Learning with Maximally Interfered Retrieval 2019 NeurIPS
Gradient based sample selection for online continual learning 2019 NeurIPS
Efficient lifelong learning with A-GEM 2019 ICLR
Task-Free Continual Learning 2019 CVPR
Continual Learning with Tiny Episodic Memories 2019 Arxiv

Online CL

In online CL, the learner is only allowed to process the data for each task once.

Paper Title Year Conference/Journal
New Insights for the Stability-Plasticity Dilemma in Online Continual Learning 2023 ICLR
Real-Time Evaluation in Online Continual Learning: A New Hope 2023 CVPR
PCR: Proxy-based Contrastive Replay for Online Class-Incremental Continual Learning 2023 CVPR
Dealing with Cross-Task Class Discrimination in Online Continual Learning 2023 CVPR
Online continual learning through mutual information maximization 2022 ICML
Online Coreset Selection for Rehearsal-based Continual Learning 2022 ICLR
New Insights on Reducing Abrupt Representation Change in Online Continual Learning 2022 ICLR
Online Continual Learning on Class Incremental Blurry Task Configuration with Anytime Inference 2022 ICLR
Information-theoretic Online Memory Selection for Continual Learning 2022 ICLR
Continual Normalization: Rethinking Batch Normalization for Online Continual Learning 2022 ICLR
Navigating Memory Construction by Global Pseudo-Task Simulation for Continual Learning 2022 NeurIPS
Not Just Selection, but Exploration: Online Class-Incremental Continual Learning via Dual View Consistency 2022 CVPR
Online Task-free Continual Learning with Dynamic Sparse Distributed Memory 2022 ECCV
Mitigating Forgetting in Online Continual Learning with Neuron Calibration 2021 NeurIPS
Online class-incremental continual learning with adversarial shapley value 2021 AAAI
Online Continual Learning with Natural Distribution Shifts: An Empirical Study with Visual Data 2021 ICCV
Continual Prototype Evolution: Learning Online from Non-Stationary Data Streams 2021 ICCV
La-MAML: Look-ahead Meta Learning for Continual Learning 2020 NeurIPS
Online Learned Continual Compression with Adaptive Quantization Modules 2020 ICML
Online Continual Learning under Extreme Memory Constraints 2020 ECCV
Online Continual Learning with Maximally Interfered Retrieval 2019 NeurIPS
Gradient based sample selection for online continual learning 2019 NeurIPS
On Tiny Episodic Memories in Continual Learning Arxiv 2019
Progress & Compress: A scalable framework for continual learning 2018 ICML

The presence of imbalanced data streams in CL (especially online CL) has drawn significant attention, primarily due to its prevalence in real-world application scenarios.

Paper Title Year Conference/Journal
Online Bias Correction for Task-Free Continual Learning 2023 ICLR
Information-theoretic Online Memory Selection for Continual Learning 2022 ICLR
SS-IL: Separated Softmax for Incremental Learning 2021 ICCV
Online Continual Learning from Imbalanced Data 2020 ICML
Maintaining Discrimination and Fairness in Class Incremental Learning 2020 CVPR
Imbalanced Continual Learning with Partitioning Reservoir Sampling 2020 ECCV
GDumb: A Simple Approach that Questions Our Progress in Continual Learning 2020 ECCV
Large scale incremental learning 2019 CVPR
IL2M: Class Incremental Learning With Dual Memory 2019 ICCV
End-to-end incremental learning 2018 ECCV

Semi-supervised CL

Semi-supervised CL is an extension of traditional CL that allows each task to incorporate unlabeled data as well.

Paper Title Year Conference/Journal
Semi-supervised drifted stream learning with short lookback 2022 SIGKDD
Ordisco: Effective and efficient usage of incremental unlabeled data for semi-supervised continual learning 2021 CVPR
Memory-Efficient Semi-Supervised Continual Learning: The World is its Own Replay Buffer 2021 IJCNN
Overcoming Catastrophic Forgetting with Unlabeled Data in the Wild 2019 ICCV

Few-shot CL

Few-shot CL refers to the scenario where a model needs to learn new tasks with only a limited number of labeled examples per task while retaining knowledge from previously encountered tasks.

Paper Title Year Conference/Journal
Warping the Space: Weight Space Rotation for Class-Incremental Few-Shot Learning 2023 ICLR
Neural Collapse Inspired Feature-Classifier Alignment for Few-Shot Class Incremental Learning 2023 ICLR
Few-Shot Class-Incremental Learning by Sampling Multi-Phase Tasks 2022 TPAMI
Dynamic Support Network for Few-Shot Class Incremental Learning 2022 TPAMI
Subspace Regularizers for Few-Shot Class Incremental Learning 2022 ICLR
MetaFSCIL: A Meta-Learning Approach for Few-Shot Class Incremental Learning 2022 CVPR
Forward Compatible Few-Shot Class-Incremental Learning 2022 CVPR
Constrained Few-shot Class-incremental Learning 2022 CVPR
Few-Shot Class-Incremental Learning via Entropy-Regularized Data-Free Replay 2022 ECCV
MgSvF: Multi-Grained Slow vs. Fast Framework for Few-Shot Class-Incremental Learning 2021 TPAMI
Semantic-aware Knowledge Distillation for Few-Shot Class-Incremental Learning 2021 CVPR
Self-Promoted Prototype Refinement for Few-Shot Class-Incremental Learning 2021 CVPR
Few-Shot Incremental Learning with Continually Evolved Classifiers 2021 CVPR
Synthesized Feature based Few-Shot Class-Incremental Learning on a Mixture of Subspaces 2021 ICCV
Few-Shot Lifelong Learning 2021 AAAI
Few-Shot Class-Incremental Learning via Relation Knowledge Distillation 2021 AAAI
Few-shot Continual Learning: a Brain-inspired Approach 2021 Arxiv
Few-Shot Class-Incremental Learning 2020 CVPR

Unsupervised CL

Unsupervised CL (UCL) assumes that only unlabeled data is provided to the CL learner.

Paper Title Year Conference/Journal
Unsupervised Continual Learning in Streaming Environments 2023 TNNLS
Representational Continuity for Unsupervised Continual Learning 2022 ICLR
Probing Representation Forgetting in Supervised and Unsupervised Continual Learning 2022 CVPR
Unsupervised Continual Learning for Gradually Varying Domains 2022 CVPRW
Co2L: Contrastive Continual Learning 2021 ICCV
Unsupervised Progressive Learning and the STAM Architecture 2021 IJCAI
Continual Unsupervised Representation Learning 2019 NeurIPS

Theoretical Analysis

Theory or analysis of continual learning

Paper Title Year Conference/Journal
The Ideal Continual Learner: An Agent That Never Forgets 2023 ICML
Continual Learning in Linear Classification on Separable Data 2023 ICML
Theory on Forgetting and Generalization of Continual Learning 2023 ArXiv
A Theoretical Study on Solving Continual Learning 2022 NeurIPS
Learning Curves for Continual Learning in Neural Networks: Self-Knowledge Transfer and Forgetting 2022 ICLR
Continual Learning in the Teacher-Student Setup: Impact of Task Similarity 2022 ICML
Formalizing the Generalization-Forgetting Trade-off in Continual Learning 2021 NeurIPS
A PAC-Bayesian Bound for Lifelong Learning 2014 ICML

Forgetting in Foundation Models

Foundation models are large machine learning models trained on a vast quantity of data at scale, such that they can be adapted to a wide range of downstream tasks.

Links: Forgetting in Fine-Tuning Foundation Models | Forgetting in One-Epoch Pre-training | CL in Foundation Model

Forgetting in Fine-Tuning Foundation Models

When fine-tuning a foundation model, there is a tendency to forget the pre-trained knowledge, resulting in sub-optimal performance on downstream tasks.

Paper Title Year Conference/Journal
Improving Gender Fairness of Pre-Trained Language Models without Catastrophic Forgetting 2023 ACL
On The Role of Forgetting in Fine-Tuning Reinforcement Learning Models 2023 ICLRW
Reinforcement Learning with Action-Free Pre-Training from Videos 2022 ICML
Video PreTraining (VPT): Learning to Act by Watching Unlabeled Online Videos 2022 NeurIPS
How Should Pre-Trained Language Models Be Fine-Tuned Towards Adversarial Robustness? 2021 NeurIPS
Mixout: Effective Regularization to Finetune Large-scale Pretrained Language Models 2020 ICLR
Recall and Learn: Fine-tuning Deep Pretrained Language Models with Less Forgetting 2020 EMNLP
Universal Language Model Fine-tuning for Text Classification 2018 ACL

Forgetting in One-Epoch Pre-training

Foundation models often undergo training on a dataset for a single pass. As a result, the earlier examples encountered during pre-training may be overwritten or forgotten by the model more quickly than the later examples.

Paper Title Year Conference/Journal
Measuring Forgetting of Memorized Training Examples 2023 ICLR
Quantifying Memorization Across Neural Language Models 2023 ICLR
Analyzing leakage of personally identifiable information in language models 2023 S&P
How Well Does Self-Supervised Pre-Training Perform with Streaming Data? 2022 ICLR
The challenges of continuous self-supervised learning 2022 ECCV
Continual contrastive learning for image classification 2022 ICME

CL in Foundation Model

By leveraging the powerful feature extraction capabilities of foundation models, researchers have been able to explore new avenues for advancing continual learning techniques.

Paper Title Year Conference/Journal
SLCA: Slow Learner with Classifier Alignment for Continual Learning on a Pre-trained Model 2023 ICCV
Progressive Prompts: Continual Learning for Language Models 2023 ICLR
Continual Pre-training of Language Models 2023 ICLR
CODA-Prompt: COntinual Decomposed Attention-based Prompting for Rehearsal-Free Continual Learning 2023 CVPR
PIVOT: Prompting for Video Continual Learning 2023 CVPR
Do Pre-trained Models Benefit Equally in Continual Learning? 2023 WACV
Revisiting Class-Incremental Learning with Pre-Trained Models: Generalizability and Adaptivity are All You Need 2023 Arxiv
First Session Adaptation: A Strong Replay-Free Baseline for Class-Incremental Learning 2023 Arxiv
Memory Efficient Continual Learning with Transformers 2022 NeurIPS
S-Prompts Learning with Pre-trained Transformers: An Occam’s Razor for Domain Incremental Learning 2022 NeurIPS
Pretrained Language Model in Continual Learning: A Comparative Study 2022 ICLR
Effect of scale on catastrophic forgetting in neural networks 2022 ICLR
Learning to Prompt for Continual Learning 2022 CVPR
Class-Incremental Learning with Strong Pre-trained Models 2022 CVPR
DualPrompt: Complementary Prompting for Rehearsal-free Continual Learning 2022 ECCV
ELLE: Efficient Lifelong Pre-training for Emerging Data 2022 ACL
Fine-tuned Language Models are Continual Learners 2022 EMNLP
Continual Training of Language Models for Few-Shot Learning 2022 EMNLP
Continual Learning with Foundation Models: An Empirical Study of Latent Replay 2022 Conference on Lifelong Learning Agents
Achieving Forgetting Prevention and Knowledge Transfer in Continual Learning 2021 NeurIPS
An Empirical Investigation of the Role of Pre-training in Lifelong Learning 2021 Arxiv

Forgetting in Domain Adaptation

The goal of domain adaptation is to transfer the knowledge from a source domain to a target domain.

Paper Title Year Conference/Journal
Continual Source-Free Unsupervised Domain Adaptation 2023 International Conference on Image Analysis and Processing
CoSDA: Continual Source-Free Domain Adaptation 2023 Arxiv
Lifelong Domain Adaptation via Consolidated Internal Distribution 2022 NeurIPS
Online Domain Adaptation for Semantic Segmentation in Ever-Changing Conditions 2022 ECCV
FRIDA -- Generative Feature Replay for Incremental Domain Adaptation 2022 CVIU
Unsupervised Continual Learning for Gradually Varying Domains 2022 CVPRW
Continual Adaptation of Visual Representations via Domain Randomization and Meta-learning 2021 CVPR
Gradient Regularized Contrastive Learning for Continual Domain Adaptation 2021 AAAI
Learning to Adapt to Evolving Domains 2020 NeurIPS
AdaGraph: Unifying Predictive and Continuous Domain Adaptation through Graphs 2019 CVPR
ACE: Adapting to Changing Environments for Semantic Segmentation 2019 ICCV
Adapting to Continuously Shifting Domains 2018 ICLRW

Forgetting in Test-Time Adaptation

Test time adaptation (TTA) refers to the process of adapting a pre-trained model on-the-fly to unlabeled test data during inference or testin.

Paper Title Year Conference/Journal
A Comprehensive Survey on Test-Time Adaptation under Distribution Shifts 2023 Arxiv
MECTA: Memory-Economic Continual Test-Time Model Adaptation 2023 ICLR
Decorate the Newcomers: Visual Domain Prompt for Continual Test Time Adaptation 2023 AAAI (Outstanding Student Paper Award)
Robust Mean Teacher for Continual and Gradual Test-Time Adaptation 2023 CVPR
A Probabilistic Framework for Lifelong Test-Time Adaptation 2023 CVPR
EcoTTA: Memory-Efficient Continual Test-time Adaptation via Self-distilled Regularization 2023 CVPR
AUTO: Adaptive Outlier Optimization for Online Test-Time OOD Detection 2023 Arxiv
Efficient Test-Time Model Adaptation without Forgetting 2022 ICML
MEMO: Test time robustness via adaptation and augmentation 2022 NeurIPS
Continual Test-Time Domain Adaptation 2022 CVPR
Improving test-time adaptation via shift-agnostic weight regularization and nearest source prototypes 2022 ECCV
Tent: Fully Test-Time Adaptation by Entropy Minimization 2021 ICLR

Forgetting in Meta-Learning

Meta-learning, also known as learning to learn, focuses on developing algorithms and models that can learn from previous learning experiences to improve their ability to learn new tasks or adapt to new domains more efficiently and effectively.

Links: Incremental Few-Shot Learning | Continual Meta-Learning

Incremental Few-Shot Learning

Incremental few-shot learning (IFSL) focuses on the challenge of learning new categories with limited labeled data while retaining knowledge about previously learned categories.

Paper Title Year Conference/Journal
Constrained Few-shot Class-incremental Learning 2022 CVPR
Meta-Learning with Less Forgetting on Large-Scale Non-Stationary Task Distributions 2022 ECCV
Overcoming Catastrophic Forgetting in Incremental Few-Shot Learning by Finding Flat Minima 2021 NeurIPS
Incremental Few-shot Learning via Vector Quantization in Deep Embedded Space 2021 ICLR
XtarNet: Learning to Extract Task-Adaptive Representation for Incremental Few-Shot Learning 2020 ICML
Incremental Few-Shot Learning with Attention Attractor Networks 2019 NeurIPS
Dynamic Few-Shot Visual Learning without Forgetting 2018 CVPR

Continual Meta-Learning

The goal of continual meta-learning (CML) is to address the challenge of forgetting in non-stationary task distributions.

Paper Title Year Conference/Journal
Adaptive Compositional Continual Meta-Learning 2023 ICML
Learning to Learn and Remember Super Long Multi-Domain Task Sequence 2022 CVPR
Meta-Learning with Less Forgetting on Large-Scale Non-Stationary Task Distributions 2022 ECCV
Variational Continual Bayesian Meta-Learning 2021 NeurIPS
Meta Learning on a Sequence of Imbalanced Domains with Difficulty Awareness 2021 ICCV
Addressing Catastrophic Forgetting in Few-Shot Problems 2020 ICML
Continuous meta-learning without tasks 2020 NeurIPS
Reconciling meta-learning and continual learning with online mixtures of tasks 2019 NeurIPS
Fast Context Adaptation via Meta-Learning 2019 ICML
Online meta-learning 2019 ICML

Forgetting in Generative Models

The goal of a generative model is to learn a generator that can generate samples from a target distribution.

Links: GAN Training is a Continual Learning Problem | Lifelong Learning of Generative Models

GAN Training is a Continual Learning Problem

Treating GAN training as a continual learning problem.

Paper Title Year Conference/Journal
Learning to Retain while Acquiring: Combating Distribution-Shift in Adversarial Data-Free Knowledge Distillation 2023 CVPR
Momentum Adversarial Distillation: Handling Large Distribution Shifts in Data-Free Knowledge Distillation 2022 NeurIPS
Robust and Resource-Efficient Data-Free Knowledge Distillation by Generative Pseudo Replay 2022 AAAI
Preventing Catastrophic Forgetting and Distribution Mismatch in Knowledge Distillation via Synthetic Data 2022 WACV
On Catastrophic Forgetting and Mode Collapse in Generative Adversarial Networks 2020 IJCNN
Generative adversarial network training is a continual learning problem 2018 ArXiv

Lifelong Learning of Generative Models

The goal is to develop generative models that can continually generate high-quality samples for both new and previously encountered tasks.

Paper Title Year Conference/Journal
The Curse of Recursion: Training on Generated Data Makes Models Forget 2023 Arxiv
Forget-Me-Not: Learning to Forget in Text-to-Image Diffusion Models 2023 Arxiv
Selective Amnesia: A Continual Learning Approach to Forgetting in Deep Generative Models 2023 Arxiv
Lifelong Generative Modelling Using Dynamic Expansion Graph Model 2022 AAAI
Continual Variational Autoencoder Learning via Online Cooperative Memorization 2022 ECCV
Hyper-LifelongGAN: Scalable Lifelong Learning for Image Conditioned Generation 2021 CVPR
Lifelong Twin Generative Adversarial Networks 2021 ICIP
Lifelong Mixture of Variational Autoencoders 2021 TNNLS
Lifelong Generative Modeling 2020 Neurocomputing
Lifelong GAN: Continual Learning for Conditional Image Generation 2019 ICCV

Forgetting in Reinforcement Learning

Reinforcement learning is a machine learning technique that allows an agent to learn how to behave in an environment by trial and error, through rewards and punishments.

Paper Title Year Conference/Journal
A Definition of Continual Reinforcement Learning 2023 Arxiv
Continual Task Allocation in Meta-Policy Network via Sparse Prompting 2023 ICML
Building a Subspace of Policies for Scalable Continual Learning 2023 ICLR
Modular Lifelong Reinforcement Learning via Neural Composition 2022 ICLR
Disentangling Transfer in Continual Reinforcement Learning 2022 NeurIPS
Towards continual reinforcement learning: A review and perspectives 2022 Journal of Artificial Intelligence Research
Model-Free Generative Replay for Lifelong Reinforcement Learning: Application to Starcraft-2 2022 Conference on Lifelong Learning Agents
Transient Non-stationarity and Generalisation in Deep Reinforcement Learning 2021 ICLR
Sharing Less is More: Lifelong Learning in Deep Networks with Selective Layer Transfer 2021 ICML
Pseudo-rehearsal: Achieving deep reinforcement learning without catastrophic forgetting 2021 Neurocomputing
Lifelong Policy Gradient Learning of Factored Policies for Faster Training Without Forgetting 2020 NeurIPS
Policy Consolidation for Continual Reinforcement Learning 2019 ICML
Exploiting Hierarchy for Learning and Transfer in KL-regularized RL 2019 Arxiv
Model-Agnostic Meta-Learning for Fast Adaptation of Deep Networks 2017 ICML
Progressive neural networks 2016 Arxiv
Learning a synaptic learning rule 1991 IJCNN

Forgetting in Federated Learning

Federated learning (FL) is a decentralized machine learning approach where the training process takes place on local devices or edge servers instead of a centralized server.

Links: Forgetting Due to Non-IID Data in FL | Federated Continual Learning

Forgetting Due to Non-IID Data in FL

This branch pertains to the forgetting problem caused by the inherent non-IID (not identically and independently distributed) data among different clients participating in FL.

Paper Title Year Conference/Journal
GradMA: A Gradient-Memory-based Accelerated Federated Learning with Alleviated Catastrophic Forgetting 2023 CVPR
Acceleration of Federated Learning with Alleviated Forgetting in Local Training 2022 ICLR
Preservation of the Global Knowledge by Not-True Distillation in Federated Learning 2022 NeurIPS
Learn from Others and Be Yourself in Heterogeneous Federated Learning 2022 CVPR
Rethinking Architecture Design for Tackling Data Heterogeneity in Federated Learning 2022 CVPR
Model-Contrastive Federated Learning 2021 CVPR
SCAFFOLD: Stochastic Controlled Averaging for Federated Learning 2020 ICML
Overcoming Forgetting in Federated Learning on Non-IID Data 2019 NeurIPSW

Federated Continual Learning

This branch addresses the issue of continual learning within each individual client in the federated learning process, which results in forgetting at the overall FL level.

Paper Title Year Conference/Journal
FedET: A Communication-Efficient Federated Class-Incremental Learning Framework Based on Enhanced Transformer 2023 IJCAI
Better Generative Replay for Continual Federated Learning 2023 ICLR
Don’t Memorize; Mimic The Past: Federated Class Incremental Learning Without Episodic Memory 2023 ICMLW
Addressing Catastrophic Forgetting in Federated Class-Continual Learning 2023 Arxiv
Federated Class-Incremental Learning 2022 CVPR
Continual Federated Learning Based on Knowledge Distillation 2022 IJCAI
Federated Continual Learning with Weighted Inter-client Transfer 2021 ICML
A distillation-based approach integrating continual learning and federated learning for pervasive services 2021 Arxiv

Beneficial Forgetting

Beneficial forgetting arises when the model contains private information that could lead to privacy breaches or when irrelevant information hinders the learning of new tasks. In these situations, forgetting becomes desirable as it helps protect privacy and facilitate efficient learning by discarding unnecessary information.

Problem Setting Goal
Mitigate Overfitting mitigate memorization of training data through selective forgetting
Debias and Forget Irrelevant Information forget biased information to achieve better performance or remove irrelevant information to learn new tasks
Machine Unlearning forget some specified training data to protect user privacy

Links: Combat Overfitting Through Forgetting | Learning New Knowledge Through Forgetting Previous Knowledge | Machine Unlearning

Forgetting Irrelevant Information to Achieve Better Performance

Combat Overfitting Through Forgetting

Overfitting in neural networks occurs when the model excessively memorizes the training data, leading to poor generalization. To address overfitting, it is necessary to selectively forget irrelevant or noisy information.

Paper Title Year Conference/Journal
Sample-Efficient Reinforcement Learning by Breaking the Replay Ratio Barrier 2023 ICLR
The Primacy Bias in Deep Reinforcement Learning 2022 ICML
Learning with Selective Forgetting 2021 IJCAI
SIGUA: Forgetting May Make Learning with Noisy Labels More Robust 2020 ICML
Invariant Representations through Adversarial Forgetting 2020 AAAI
Forget a Bit to Learn Better: Soft Forgetting for CTC-based Automatic Speech Recognition 2019 Interspeech

Learning New Knowledge Through Forgetting Previous Knowledge

"Learning to forget" suggests that not all previously acquired prior knowledge is helpful for learning new tasks.

Paper Title Year Conference/Journal
ReFactor GNNs: Revisiting Factorisation-based Models from a Message-Passing Perspective 2022 NeurIPS
Fortuitous Forgetting in Connectionist Networks 2022 ICLR
Skin Deep Unlearning: Artefact and Instrument Debiasing in the Context of Melanoma Classification 2022 ICML
Near-Optimal Task Selection for Meta-Learning with Mutual Information and Online Variational Bayesian Unlearning 2022 AISTATS
AFEC: Active Forgetting of Negative Transfer in Continual Learning 2021 NeurIPS
Active Forgetting: Adaptation of Memory by Prefrontal Control 2021 Annual Review of Psychology
Learning to Forget for Meta-Learning 2020 CVPR
The Forgotten Part of Memory 2019 Nature
Learning Not to Learn: Training Deep Neural Networks with Biased Data 2019 CVPR
Inhibiting your native language: the role of retrieval-induced forgetting during second-language acquisition 2007 Psychological Science

Machine Unlearning

Machine unlearning, a recent area of research, addresses the need to forget previously learned training data in order to protect user data privacy.

Paper Title Year Conference/Journal
Deep Unlearning via Randomized Conditionally Independent Hessians 2022 CVPR
Eternal Sunshine of the Spotless Net: Selective Forgetting in Deep Networks 2022 CVPR
PUMA: Performance Unchanged Model Augmentation for Training Data Removal 2022 AAAI
ARCANE: An Efficient Architecture for Exact Machine Unlearning 2022 IJCAI
Learn to Forget: Machine Unlearning via Neuron Masking 2022 IEEE TDSC
Backdoor Defense with Machine Unlearning 2022 IEEE INFOCOM
Markov Chain Monte Carlo-Based Machine Unlearning: Unlearning What Needs to be Forgotten 2022 ASIA CCS
Machine Unlearning 2021 SSP
Remember What You Want to Forget: Algorithms for Machine Unlearning 2021 NeurIPS
Machine Unlearning via Algorithmic Stability 2021 COLT
Variational Bayesian Unlearning 2020 NeurIPS
Rapid retraining of machine learning models 2020 ICML
Certified Data Removal from Machine Learning Models 2020 ICML
Making AI Forget You: Data Deletion in Machine Learning 2019 NeurIPS
Lifelong Anomaly Detection Through Unlearning 2019 CCS
The EU Proposal for a General Data Protection Regulation and the Roots of the ‘Right to Be Forgotten’ 2013 Computer Law & Security Review

Contact

We welcome all researchers to contribute to this repository 'forgetting in deep learning'.

Email: wangzhenyineu@gmail.com | ennengyang@stumail.neu.edu.cn