The paper addresses the problem of continual test-time adaptation, where the goal is to adapt a pre-trained model to a sequence of unlabelled target domains at test time. Existing methods on test-time training suffer from several limitations: (1) Mismatch between the feature extractor and classifier; (2) Interference between the main and self-supervised tasks; (3) Lack of the ability to quickly adapt to the current distribution.
To address these challenges, the authors propose a cascading paradigm that simultaneously updates the feature extractor and classifier at test time, mitigating the mismatch between them and enabling long-term model adaptation. The pre-training of the model is structured within a meta-learning framework, thereby minimizing the interference between the main and self-supervised tasks and encouraging fast adaptation in the presence of limited unlabelled data.
Additionally, the authors introduce two new evaluation metrics - average accuracy and forward transfer - to effectively measure the model's adaptation capabilities in dynamic, real-world scenarios. Extensive experiments and ablation studies demonstrate the superiority of the proposed approach in a range of tasks including image classification, text classification, and speech recognition.
Sang ngôn ngữ khác
từ nội dung nguồn
arxiv.org
Thông tin chi tiết chính được chắt lọc từ
by Kien X. Nguy... lúc arxiv.org 10-03-2024
https://arxiv.org/pdf/2407.12240.pdfYêu cầu sâu hơn