View a PDF of the paper titled Prototype Augmented Hypernetworks for Continual Learning, by Neil De La Fuente and 6 other authors
View PDF
HTML (experimental)
Abstract:Continual learning (CL) aims to learn a sequence of tasks without forgetting prior knowledge, but gradient updates for a new task often overwrite the weights learned earlier, causing catastrophic forgetting (CF). We propose Prototype-Augmented Hypernetworks (PAH), a framework where a single hypernetwork, conditioned on learnable task prototypes, dynamically generates task-specific classifier heads on demand. To mitigate forgetting, PAH combines cross-entropy with dual distillation losses, one to align logits and another to align prototypes, ensuring stable feature representations across tasks. Evaluations on Split-CIFAR100 and TinyImageNet demonstrate that PAH achieves state-of-the-art performance, reaching 74.5 % and 63.7 % accuracy with only 1.7 % and 4.4 % forgetting, respectively, surpassing prior methods without storing samples or heads.
Submission history
From: Neil De La Fuente [view email]
[v1]
Mon, 12 May 2025 11:25:54 UTC (801 KB)
[v2]
Tue, 13 May 2025 07:08:25 UTC (612 KB)