Core Concepts
We propose two novel algorithms, NEURONAL-S and NEURONAL-P, for stream-based and pool-based active learning that mitigate the adverse impacts of the number of classes (K) on the performance and computational costs of existing bandit-based approaches. Our methods achieve improved theoretical guarantees and empirical performance.
Abstract
The paper studies both stream-based and pool-based active learning with neural network approximations. Recent works have proposed bandit-based approaches that transform active learning into a bandit problem, achieving theoretical and empirical success. However, the performance and computational costs of these methods may be susceptible to the number of classes (K) due to this transformation.
To address this issue, the authors propose two algorithms:
-
NEURONAL-S (stream-based):
- Redesigns the input and output of the exploitation and exploration neural networks to directly take the d-dimensional instance as input and output the predicted probabilities for K classes synchronously, mitigating the curse of K.
- Introduces an end-to-end embedding as the input of the exploration network to remove the dependence on the input dimension while preserving essential information.
- Provides theoretical regret bounds that grow slower than existing bandit-based approaches concerning K.
-
NEURONAL-P (pool-based):
- Brings the redesigned exploitation and exploration networks into the pool-based setting.
- Proposes a novel gap-inverse-based selection strategy tailored for pool-based active learning.
- Provides a performance analysis in the non-parametric setting for neural network models.
The authors conduct extensive experiments on various datasets, demonstrating that the proposed algorithms consistently outperform state-of-the-art baselines in both stream-based and pool-based settings.
Stats
The number of classes (K) ranges from 2 to 10 across the datasets.
The authors evaluate the test accuracy and running time of the proposed algorithms and baselines.
Quotes
"We study both stream-based and pool-based active learning with neural network approximations."
"We propose two algorithms based on the newly designed exploitation and exploration neural networks for stream-based and pool-based active learning."
"We provide theoretical performance guarantees for both algorithms in a non-parametric setting, demonstrating a slower error-growth rate concerning K for the proposed approaches."