核心概念
The author argues that the decline in active learning performance using proxy-based methods stems from selecting redundant samples and compromising valuable pre-trained information. The proposed ASVP method aligns features and training to improve efficiency.
要約
The paper discusses the challenges of efficient active learning with pre-trained models and introduces the ASVP method to address performance issues. It analyzes factors affecting active learning performance, proposes solutions, and validates the effectiveness through experiments on various datasets.
The study highlights the importance of preserving valuable pre-trained information and reducing redundant sample selection in active learning. By aligning features and training methods, the ASVP method improves efficiency while maintaining computational speed. The proposed sample savings ratio metric provides a clear measure of cost-effectiveness in active learning strategies.
統計
Recent benchmarks have demonstrated uncertainty sampling as a state-of-the-art approach when fine-tuning pre-trained models.
The accuracy achieved by a model trained with labeled samples is used to estimate the number of samples required for a random baseline.
The total cost of active learning includes annotation costs and training costs based on AWS EC2 instances.
Experiments validate that ASVP consistently outperforms SVPp in terms of sample saving ratio and overall cost reduction.
引用
"Practitioners therefore have to face the hassle of weighing computational efficiency against overall cost."
"Our method significantly improves the total cost of efficient active learning while maintaining computational efficiency."