Core Concepts
Mimicking the representations of the oracle model, which is trained on all classes, at the initial phase of class incremental learning can significantly boost the overall performance.
Abstract
The paper investigates improving class incremental learning (CIL) by focusing on the initial phase, which is often overlooked in previous works.
Key highlights:
The authors find that directly encouraging the CIL learner to output similar representations as the oracle model (trained on all classes) at the initial phase can greatly boost the CIL performance.
Through eigenvalue analysis, the authors discover that compared to the na??vely-trained initial-phase model, the data representations of each class produced by the oracle model scatter more uniformly.
Inspired by this observation, the authors propose a novel Class-wise Decorrelation (CwD) regularization technique to enforce representations of each class to scatter more uniformly at the initial CIL phase.
Extensive experiments show that CwD consistently and significantly improves the performance of existing state-of-the-art CIL methods by around 1% to 3%.
The authors also conduct detailed ablation studies to understand the impact of factors like the number of classes at the initial phase, the number of exemplars, and the CwD regularization coefficient.
Stats
The paper does not contain any key metrics or important figures to support the author's key logics.
Quotes
The paper does not contain any striking quotes supporting the author's key logics.