EVCL: Elastic Variational Continual Learning with Weight Consolidation

Autor: Batra, Hunar, Clark, Ronald
Rok vydání: 2024
Předmět:
Druh dokumentu: Working Paper
Popis: Continual learning aims to allow models to learn new tasks without forgetting what has been learned before. This work introduces Elastic Variational Continual Learning with Weight Consolidation (EVCL), a novel hybrid model that integrates the variational posterior approximation mechanism of Variational Continual Learning (VCL) with the regularization-based parameter-protection strategy of Elastic Weight Consolidation (EWC). By combining the strengths of both methods, EVCL effectively mitigates catastrophic forgetting and enables better capture of dependencies between model parameters and task-specific data. Evaluated on five discriminative tasks, EVCL consistently outperforms existing baselines in both domain-incremental and task-incremental learning scenarios for deep discriminative models.
Comment: Accepted at ICML 2024 Workshop on Structured Probabilistic Inference & Generative Modeling
Databáze: arXiv