Abstract
The replay-based approaches are a notable family of methods among many efforts on Continual Learning, where memory sampling strat- egy and rehearsal mode are two fundamental aspects to alleviate the catastrophic forgetting. However, most existing replay-based approaches focus primarily on exploring the rehearsal mode but neglect the sig- nificant influence of the sampling strategy, especially failing to ade- quately utilize the inherent attributes in samples and the information provided by the old task model. To this end, we propose a novel sam- pling strategy dubbed Uncertainty-Aware Sampling (UAS) strategy, which employs model and data uncertainties as criteria to select sam- ples that are stable to the model and have low noise for rehearsal. Further, we design a dual network to acquire new knowledge while maintaining old knowledge, in which a Convolutional Neural Network (CNN) is applied to continuously learn and consolidate knowledge, and a Bayesian Neural Network (BNN) is employed as a comple- ment to capture the uncertainty while providing additional information for the CNN. Besides, we incorporate a data uncertainty loss into Dark Experience Replay as rehearsal mode to alleviate the catas- trophic forgetting in both CNN and BNN, called Uncertainty-Aware Replay (UAR). Extensive experiments on four benchmark datasets demonstrate that the proposed framework is competitive with state- of-the-art methods under three different continual learning settings.
Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.Data availability
The datasets analysed during the current study are available from the corresponding author on reasonable request.
Code availability
The code for this article will be available after the article is accepted or from the corresponding author on reasonable request.
References
De Lange M, Aljundi R, Masana M, Parisot S, Jia X, Leonardis A, Slabaugh G, Tuytelaars T (2022) A continual learning survey: Defying forgetting in classification tasks. IEEE Trans Pattern Anal Mach Intell 44(7):3366–3385
Zhao K, Fu Z, Yang J (2023) Continual learning via region-aware memory. Appl Intell 53:8389–8401
Fu Y, Cao H, Chen X, Ding J (2022) Task-incremental broad learn- ing system for multi-component intelligent fault diagnosis of machinery. Knowl-Based Syst 246:108730–108744
Li D, Liu S, Gao F, Sun X (2022) Continual learning classification method and its application to equipment fault diagnosis. Appl Intell 52:858–874
Li D, Gu M, Liu S, Sun X, Gong L, Qian K (2022) Continual learning classification method with the weighted k-nearest neighbor rule for time- varying data space based on the artificial immune system. Knowl-Based Syst 240:108145–108160
McCloskey M, Cohen NJ (1989) Catastrophic interference in connectionist networks: The sequential learning problem. Psychol Learn Motiv 24:109–165
Jiang M, Li F, Li L (2022) Continual meta-learning algorithm. Appl Intell 52:4527–4542
Jonathan S, Wojciech C, Jelena L, Agnieszka G-B, Yee WT, Razvan P, Raia H (2018) Progress & compress: a scalable framework for continual learning. In: International conference on machine learning pp 4528–4537
Li Z, Hoiem D (2017) Learning without forgetting. Transactions on pattern analysis and machine intelligence. IEEE 40(12):2935–2947
Zenke F, Poole B, Ganguli S (2017) Continual learning through synaptic intelligence. In: International conference on machine learning pp 3987–3995
Rusu AA, Rabinowitz NC, Desjardins G, Soyer H, Kirkpatrick J, Kavukcuoglu K, Pascanu R, Hadsell R (2016) Progressive neural networks. ArXiv preprint. https://arxiv.org/abs/1606.04671
Serra J, Suris D, Miron M, Karatzoglou A(2018) Overcoming catastrophic forgetting with hard attention to the task. In: International conference on machine learning pp 4548–4557
Ke Z, Liu B, Huang X (2020) Continual learning of a mixed sequence of similar and dissimilar tasks. In: Advances in neural information processing systems pp 18493–18504
Dekhovich A, Tax DMJ, Sluiter MHF, Bessa MA (2023) Continual prune-and-select: class-incremental learning with specialized subnetworks. Appl Intell 1–18
Abati D, Tomczak JM, Blankevoort T, Calderara S, Cucchiara R, Bejnordi BE (2020) Conditional channel gated networks for task-aware continual learning. In: Proceedings of the IEEE conference on computer vision and pattern recognition pp 3931–3940
Aljundi R, Lin M, Goujaud B, Bengio Y (2019) Gradient based sample selection for online continual learning. In: Advances in neural information processing systems pp 11816–11825
Chaudhry A, Ranzato MA, Rohrbach M, Elhoseiny M (2019) Efficient lifelong learning with A-GEM. In: International conference on learning representations pp 1–20
Rebuffi S-A, Kolesnikov A, Sperl G, Lampert CH (2017) Icarl: incremental classifier and representation learning. In: Proceedings of the IEEE conference on computer vision and pattern recognition pp 2001–2010
Benjamin A, Rolnick D, Kording K (2018) Measuring and regularizing networks in function space. In: International conference on learning representations pp 1–18
Chaudhry A, Gordo A, Dokania PK, Torr P, Lopez-Paz D (2020) Using hindsight to anchor past knowledge in continual learning. arXiv preprint arXiv:2002.08165
Buzzega P, Boschini M, Porrello A, Abati D, Calderara S (2020) Dark experience for general continual learning: a strong, simple baseline. In: Advances in neural information processing systems pp 1–24
Ji Z, Liu J, Wang Q, Zhang Z (2021) Coordinating experience replay: A harmonious experience retention approach for continual learning. Knowl-Based Syst 234:107589–107601
Buzzega P, Boschini M, Porrello A, Calderara S (2021) Rethinking experience replay: a bag of tricks for continual learning. In: Proceedings of the international conference on pattern recognition pp 2180–2187
Quang P, Chenghao L, Steven H (2021) DualNet: continual learning, fast and slow. In: Proceedings of the advances in neural information processing systems pp 16131–16144
Simon C, Koniusz P, Harandi M (2021) On learning the geodesic path for incremental learning. In: Proceedings of the IEEE conference on computer vision and pattern recognition pp 1591–1600
Aljundi R, Belilovsky E, Tuytelaars T, Charlin L, Caccia M, Lin M, Page-Caccia L (2019) Online continual learning with maximal interfered retrieval. In: Proceedings of the advances in neural information processing systems pp 1–15
Gal Y (2016) Uncertainty in deep learning. PhD thesis, University of Cambridge
Mobiny A, Yuan P, Moulik SK, Garg N, Wu CC, Van Nguyen H (2021) Dropconnect is effective in modeling uncertainty of bayesian deep networks. Sci Rep 11(1):1–14
Robins A (1995) Catastrophic forgetting, rehearsal and pseudorehearsal. Con- nection Science 7(2):123–146
MacKay DJC (1992) A practical bayesian framework for backpropagation networks. Neural Comput 4(3):448–472
Blundell C, Cornebise J, Kavukcuoglu K, Wierstra D (2015) Weight uncertainty in neural network. In: International conference on machine learning pp 1613–1622
Hernandez-Lobato J, Li Y, Rowland M, Bui T, Hernández-Lobato D, Turner R (2016) Black-box alpha divergence minimization. In: International conference on machine learning pp 1511–1520
Zhang Z, Lan C, Zeng W, Chen Z, Chang S-F (2020) Uncertainty-aware few-shot image classification. Int Joint Conf Artif Intell pp 1–9
Mukherjee S, Awadallah AH (2019) Uncertainty-aware self-training for text classification with few labels. In: Advances in neural information processing systems pp 1–14
Nguyen CV, Li Y, Bui TD, Turner RE (2018) Variational continual learning. In: International conference on learning representations pp 1–18
Kochurov M, Garipov T, Podoprikhin D, Molchanov D, Ashukha A, Vetrov DP (2018) Bayesian incremental learning for deep neural networks. ArXiv preprint. https://arxiv.org/abs/1802.07329
Kurmi VK, Patro BN, Subramanian VK, Namboodiri VP (2021) Do not forget to attend to uncertainty while mitigating catastrophic forgetting. In: 2021 IEEE winter conference on applications of computer vision pp 736–745
Ebrahimi S, Elhoseiny M, Darrell T, Rohrbach M (2020) Uncertainty-guided continual learning with bayesian neural networks. In: International conference on learning representations pp 1–16
Srivastava N, Hinton GE, Krizhevsky A, Sutskever I, Salakhutdinov R (2014) Dropout: a simple way to prevent neural networks from overfitting. J Mach Learn Res 1–30
Houlsby N, Huszár F, Ghahramani Z, Lengyel M (2011) Bayesian active learning for classification and preference learning. ArXiv preprint. https://arxiv.org/abs/1112.5745
Gal Y, Islam R, Ghahramani Z (2017) Deep bayesian active learning with image data. In: International conference on machine learning pp 1183–1192
Van de Ven GM, Tolias AS (2019) Three scenarios for continual learning. ArXiv preprint. https://arxiv.org/abs/1904.07734
Krizhevsky A, Hinton G et al (2009) Learning multiple layers of features from tiny images. Citeseer
Stanford (2015) Tiny imagenet challenge (CS231n). https://tiny-imagenet.herokuapp.com
Delange M, Aljundi R, Masana M, Parisot S, Jia X, Leonardis A, Slabaugh G, Tuytelaars T (2021) Continual learning: A comparative study on how to defy forgetting in classification tasks. IEEE Trans Pattern Anal Mach Intell 42(3):99
He KM, Zhang XY, Ren SQ, Sun J (2016) Deep residual learning for image recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition pp 770–778
Cha H, Lee J, Shin J (2021) Co2L: contrastive continual learning. In: Proceedings of the IEEE/CVF International conference on computer vision (ICCV) pp 9516–9525
Vitter JS (1985) Random sampling with a reservoir. ACM Transactions on Mathematical Software 11(1):37–57
Funding
This work was supported by the National Key Research and Development Program of China (Grant No. 2022ZD0160403) and the National Natural Science Foundation of China (Grant No. 62176178).
Author information
Authors and Affiliations
Contributions
All authors contributed to the study concep- tion and design. Qiang Wang: Methodology, Writing, Software. Zhong Ji: Conceptualization, Writing, Funding acquisition. Yanwei Pang: Conceptual- ization, Writing review and editing. Zhongfei Zhang: Methodology, Writing review and editing.
Corresponding author
Ethics declarations
Conflict of interest
The authors have no competing interests to declare that are relevant to the content of this article.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Wang, Q., Ji, Z., Pang, Y. et al. Uncertainty-aware enhanced dark experience replay for continual learning. Appl Intell 54, 7135–7150 (2024). https://doi.org/10.1007/s10489-024-05488-w
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10489-024-05488-w