toplogo
Sign In

Active Generalized Category Discovery: Addressing Inherent Challenges in GCD with AGCD


Core Concepts
The author proposes Active Generalized Category Discovery (AGCD) to address the inherent challenges of Generalized Category Discovery (GCD) by actively selecting valuable samples for labeling from the oracle. The approach aims to improve the performance of GCD by considering novelty, informativeness, and diversity in sample selection.
Abstract
The paper introduces AGCD as a solution to the challenges faced in GCD. By actively selecting samples for labeling and proposing an adaptive sampling strategy, the method aims to enhance category discovery performance. Experiments show that AGCD achieves state-of-the-art results on various datasets by addressing imbalanced accuracy and confidence issues between old and new classes. Key points: Introduction of Active Generalized Category Discovery (AGCD) Proposal of an adaptive sampling strategy for sample selection State-of-the-art performance achieved through experiments on different datasets In essence, AGCD is designed to improve upon existing methods by actively selecting samples for labeling based on novelty, informativeness, and diversity considerations.
Stats
Our method improves the new accuracy of GCD by 25.52%/23.49% on CUB/Air with only ∼ 2.5 samples labeled per class. The proposed Adaptive-Novel strategy consistently outperforms other query strategies across various settings.
Quotes

Key Insights Distilled From

by Shijie Ma,Fe... at arxiv.org 03-08-2024

https://arxiv.org/pdf/2403.04272.pdf
Active Generalized Category Discovery

Deeper Inquiries

How does AGCD compare to traditional active learning methods?

AGCD differs from traditional active learning methods in several key aspects: Open-World Setting: AGCD is an open-world extrapolated version of active learning, where the unlabeled data contains novel categories beyond what is present in the labeled data. This contrasts with traditional active learning, which typically operates in a closed-world setting where labeled and unlabeled classes are identical. Novelty Consideration: In AGCD, models actively select samples from both old and new classes for labeling, prioritizing novel samples for annotation. Traditional active learning methods may not consider this aspect and focus solely on uncertainty or diversity metrics. Cluster Nature of GCD: AGCD addresses the clustering nature of Generalized Category Discovery (GCD) by adaptively selecting informative novel samples at different stages of training to stabilize clusters and improve performance. Balanced Performance: AGCD aims to achieve more balanced accuracy between old and new classes compared to traditional active learning methods that may struggle with imbalanced classification performance when faced with completely unlabeled new classes. Overall, AGCD extends the principles of active learning to handle scenarios where there are unknown categories present in the data, requiring models to learn from both familiar and unfamiliar classes simultaneously.

How can EMA be applied in model training for AGCD?

Exponential Moving Average (EMA) plays a crucial role in enhancing model stability and consistency during Active Generalized Category Discovery (AGCD). Here's how EMA can be effectively utilized: Consistent Label Mapping: EMA helps maintain a stable label mapping function between ground truth labels and model predictions across different rounds of training. This ensures that queried labels can be accurately translated into the model's label space even when dealing with varying class distributions or changing cluster configurations. Inductive Evaluation Support: By using EMA-based predictions for label mapping on limited labeled data available at each round, models can provide consistent results during evaluation on unseen test datasets without access to ground truth labels. Improved Training Stability: The use of EMA reduces fluctuations in label mappings over epochs or rounds, leading to more reliable updates during training iterations while minimizing errors caused by inconsistent labeling information. In essence, incorporating EMA into model training for AG...
0
visual_icon
generate_icon
translate_icon
scholar_search_icon
star