JOURNAL ARTICLE

Variational Density Propagation Continual Learning

Abstract

Deep Neural Networks (DNNs) deployed to the real world are regularly subject to out-of-distribution (OoD) data, various types of noise, and shifting conceptual objectives. This paper proposes a framework for adapting to data distribution drift modeled by benchmark Continual Learning datasets. We develop and evaluate a method of Continual Learning that leverages uncertainty quantification from Bayesian Inference to mitigate catastrophic forgetting. We expand on previous approaches by removing the need for Monte Carlo sampling of the model weights to sample the predictive distribution. We optimize a closed-form Evidence Lower Bound (ELBO) objective approximating the predictive distribution by propagating the first two moments of a distribution, i.e. mean and covariance, through all network layers. Catastrophic forgetting is mitigated by using the closed-form ELBO to approximate the Minimum Description Length (MDL) Principle, inherently penalizing changes in the model likelihood by minimizing the KL Divergence between the variational posterior for the current task and the previous task's variational posterior acting as the prior. Leveraging the approximation of the MDL principle, we aim to initially learn a sparse variational posterior and then minimize additional model complexity learned for subsequent tasks. Our approach is evaluated for the task incremental learning scenario using density propagated versions of fully-connected and convolutional neural networks across multiple sequential benchmark datasets with varying task sequence lengths. Ultimately, this procedure produces a minimally complex network over a series of tasks mitigating catastrophic forgetting.

Keywords:
Forgetting Computer science Minimum description length Benchmark (surveying) Artificial intelligence Machine learning Divergence (linguistics) Posterior probability Inference Bayesian inference Artificial neural network Task (project management) Bayesian probability Algorithm

Metrics

0
Cited By
0.00
FWCI (Field Weighted Citation Impact)
25
Refs
0.14
Citation Normalized Percentile
Is in top 1%
Is in top 10%

Topics

Data Stream Mining Techniques
Physical Sciences →  Computer Science →  Artificial Intelligence
Domain Adaptation and Few-Shot Learning
Physical Sciences →  Computer Science →  Artificial Intelligence
Machine Learning and Data Classification
Physical Sciences →  Computer Science →  Artificial Intelligence

Related Documents

JOURNAL ARTICLE

Variational continual learning

Cuong V. NguyenYingzhen LiThang D. BuiRichard E. Turner

Journal:   Apollo (University of Cambridge) Year: 2019
JOURNAL ARTICLE

Unsupervised Generative Variational Continual Learning

Liu GuimengYang GuoCheryl Wong Sze YinPonnuthurai Nagartnam SuganathanSavitha Ramasamy

Journal:   2022 IEEE International Conference on Image Processing (ICIP) Year: 2022 Pages: 4028-4032
JOURNAL ARTICLE

Variational Data-Free Knowledge Distillation for Continual Learning

Xiaorong LiShipeng WangJian SunZongben Xu

Journal:   IEEE Transactions on Pattern Analysis and Machine Intelligence Year: 2023 Vol: 45 (10)Pages: 12618-12634
© 2026 ScienceGate Book Chapters — All rights reserved.