A Data-Free, Spectral Solution to Catastrophic Forgetting in Continual Learning.
Anastrophic Regularization (
Traditional methods like Elastic Weight Consolidation (EWC) attempt to prevent forgetting by anchoring individual parameters using the diagonal of the Fisher Information Matrix. This over-constrains the network and limits its plasticity for new tasks.
- Maximum Plasticity: Individual weights are free to change as long as the global periodic functional invariants of the previous task are maintained.
- 100% Data-Free & Privacy Preserving: EWC requires computing gradients over old datasets, posing significant privacy and storage issues.
$\mathcal{R}_{ana}$ operates purely in the spectral domain, requiring absolutely zero old data. - Computationally Efficient: It leverages fast Fourier transformations (FFTs) to seamlessly minimize harmonic frustration without needing complex state-space reconstructions.
The regularizer guides weights along Fisher-Rao geodetic paths by optimizing two key spectral metrics:
- Spectral Coherence (
$\Phi$ ): A geometric measure of phase alignment on the spectral torus. - Anastrophic Beta (
$BB$ ): Measures the harmonic tension between current and previous weight matrices.
See main.py for a plug-and-play PyTorch implementation demonstrating
Read the full theoretical framework on Zenodo: [https://zenodo.org/records/18699347]