核心概念
Proposing a Bi-Level Active Finetuning framework to balance diversity and uncertainty in sample selection.
摘要
The article introduces the Bi-Level Active Finetuning Framework (BiLAF) to address challenges in active learning methods for sample selection. It focuses on balancing diversity and uncertainty by selecting central samples for diversity and boundary samples for uncertainty. The framework operates in two stages: Core Samples Selection and Boundary Samples Selection. The process starts with identifying pseudo-class centers, followed by denoising methods and iterative strategies for boundary sample selection without relying on ground-truth labels. Extensive experiments demonstrate the efficacy of the method, outperforming existing baselines significantly. The article also discusses related work, decision boundaries in neural networks, and the importance of boundary samples in classification tasks.
統計資料
Our method achieves a remarkable improvement of nearly 3% on CIFAR100 and approximately 1% on ImageNet.
We set the core number K as 50(0.1%), 250(0.5%), 6405(0.5%) for CIFAR10, CIFAR100, and ImageNet separately.
In the boundary samples selection stage, we set nearest neighbors number k as 10, both removal ratio Prm and clustering fraction Pin as 10%, opponent penalty coefficient δ as 1.1.
For all three datasets, we resize images to 224 × 224 consistent with the pretraining for both data selection and supervised finetuning.
引述
"Our comprehensive experiments provide both qualitative and quantitative evidence of our method’s efficacy."
"Our approach is evaluated using three widely recognized datasets: CIFAR10, CIFAR100, and ImageNet-1k."
"Our objective is to choose the optimal sampling strategy to select the labeled set under the given budget to minimize the expectation error of the finetuned model."