Core Concepts
Bait, a deep active learning strategy based on Fisher Information, demonstrates superior performance but suffers from high computational and memory requirements. This work introduces two approximation methods, Bait (Exp) and Bait (Binary), to enhance Bait's computational efficiency and scalability across large-scale image datasets without compromising performance.
Abstract
This paper introduces two methods to enhance the computational efficiency and scalability of the Bait deep active learning strategy. Bait is a recently proposed approach that leverages the Fisher Information Matrix (FIM) to optimize the Bayes risk and select informative instances for training deep neural networks. However, the high computational and memory requirements of Bait have hindered its applicability on large-scale classification tasks.
The first approximation, Bait (Exp), reduces the time complexity of computing the FIM by focusing on the expectation over the most probable classes instead of the full categorical distribution. This effectively decreases the cubic dependency on the number of classes to a quadratic one.
The second approximation, Bait (Binary), aims to decouple the time and space complexity from the number of classes by reformulating the multi-class classification task into a binary one. This significantly reduces the dimensionality of the gradients and the FIM, leading to a time complexity that is independent of the number of classes.
The authors provide a comprehensive evaluation of their approximations across a variety of image datasets, ranging from small to large-scale, and compare them to other state-of-the-art deep active learning strategies. The results demonstrate that the proposed approximations achieve strong performance while considerably reducing the computational requirements, enabling the efficient use of Bait on large-scale datasets like ImageNet.
Furthermore, the authors provide an open-source toolbox that implements recent state-of-the-art deep active learning strategies, including their versions of Bait, to facilitate the adoption of these methods in future research.
Stats
The authors report the following key metrics:
Accuracy improvement over random instance selection (AUC) for various datasets
Acquisition time per cycle (in seconds) on both CPU and GPU
Quotes
"Bait's high computational and memory requirements hinder its applicability on large-scale classification tasks, resulting in current research neglecting Bait in their evaluation."
"Our unified and comprehensive evaluation across a variety of datasets demonstrates that our approximations achieve strong performance with considerably reduced time complexity."