核心概念
An amortized active learning method is proposed to efficiently select informative data points for learning nonparametric functions, without repeated model training and acquisition optimization.
摘要
The paper presents an amortized active learning (AL) approach for nonparametric function regression tasks. The key idea is to decouple the model training and acquisition function optimization from the AL loop, which can be computationally expensive, especially for nonparametric models like Gaussian processes (GPs).
The authors propose to train a neural network (NN) policy that can directly suggest informative data points for labeling, without the need for costly model training and acquisition optimization at each AL iteration. The NN policy is trained in a simulated AL environment, where GP functions are sampled, and the policy is optimized to maximize the entropy or a regularized entropy objective.
The training pipeline involves:
- Sampling GP functions and noise realizations to construct a rich distribution of nonparametric functions.
- Simulating AL experiments on the sampled functions, where the NN policy selects data points.
- Optimizing the NN policy to maximize the entropy or regularized entropy of the selected data points.
This amortized approach avoids the cubic time complexity of GP modeling and acquisition optimization, enabling real-time data selection during AL deployment. The authors demonstrate the effectiveness of their method on several benchmark regression tasks, showing that the amortized AL approach can achieve comparable performance to the time-consuming baseline GP AL method, while being significantly faster in the data selection process.
统计
The time complexity of the conventional GP AL method is O((Ninit + t-1)^3) at each iteration t, where Ninit is the initial dataset size.
The time complexity of the amortized AL deployment (NN policy forwarding) is O((Ninit + t-1)^2) at each iteration t.
引用
"Active learning (AL) is a sequential learning scheme aiming to reduce the effort and cost of labeling data [1–3]."
"To perform AL, however, one would face multiple challenges: (i) training models for every query can be nontrivial, especially when the learning time is constrained [4–6]; (ii) acquisition criteria need to be selected a priori but none of them clearly outperforms the others in all cases, which makes the selection difficult [7, 8]; (iii) optimizing an acquisition function can be difficult (e.g. sophisticated discrete search space [9])."