The paper introduces a new continual learning setting that accounts for the delay between the arrival of new data and the corresponding labels. In this setting, at each time step, the model is presented with a batch of unlabeled data from the current time step and labeled data from a previous time step, with a delay of d steps.
The authors first analyze the performance of a naive approach that only trains on the delayed labeled data, ignoring the unlabeled data. They find that the performance of this approach degrades significantly as the delay increases. The authors then explore several paradigms to leverage the unlabeled data, including semi-supervised learning via pseudo-labeling, self-supervised semi-supervised learning, and test-time adaptation. However, they find that none of these methods are able to outperform the naive baseline under the same computational budget.
To address this challenge, the authors propose a novel method called Importance Weighted Memory Sampling (IWMS). IWMS selectively rehearses labeled samples from a memory buffer that are most similar to the current unlabeled data, allowing the model to effectively adapt to the newer data distribution despite the label delay. The authors show that IWMS consistently outperforms the naive baseline and other methods across various delay and computational budget scenarios, often recovering a significant portion of the accuracy gap caused by the label delay.
The paper provides a comprehensive analysis of the impact of label delay on continual learning performance, highlighting the importance of addressing this challenge in real-world applications. The proposed IWMS method offers a simple yet effective solution that can be easily integrated into existing continual learning frameworks.
他の言語に翻訳
原文コンテンツから
arxiv.org
深掘り質問