The paper addresses the problem of continual test-time adaptation, where the goal is to adapt a pre-trained model to a sequence of unlabelled target domains at test time. Existing methods on test-time training suffer from several limitations: (1) Mismatch between the feature extractor and classifier; (2) Interference between the main and self-supervised tasks; (3) Lack of the ability to quickly adapt to the current distribution.
To address these challenges, the authors propose a cascading paradigm that simultaneously updates the feature extractor and classifier at test time, mitigating the mismatch between them and enabling long-term model adaptation. The pre-training of the model is structured within a meta-learning framework, thereby minimizing the interference between the main and self-supervised tasks and encouraging fast adaptation in the presence of limited unlabelled data.
Additionally, the authors introduce two new evaluation metrics - average accuracy and forward transfer - to effectively measure the model's adaptation capabilities in dynamic, real-world scenarios. Extensive experiments and ablation studies demonstrate the superiority of the proposed approach in a range of tasks including image classification, text classification, and speech recognition.
Till ett annat språk
från källinnehåll
arxiv.org
Djupare frågor