核心概念
Introducing Reduced Labels for Long-Tailed Data to preserve supervised information and decrease labeling costs.
摘要
The content discusses the challenges of long-tailed data in classification tasks, introduces Reduced Labels as a solution, and presents an unbiased framework for learning from these labels. It includes an analysis of the proposed method's effectiveness, data augmentation techniques used, and estimation error bounds. Experimental results on various datasets are provided to validate the approach.
Abstract:
- Long-tailed data poses challenges in classification tasks.
- Reduced Labels introduced to mitigate labeling costs and preserve supervised information.
- Unbiased framework developed for learning from Reduced Labels.
Introduction:
- Deep neural networks face limitations with long-tailed datasets.
- Precise annotation is required for long-tailed datasets.
- Weakly supervised learning methods aim to reduce labeling costs.
Proposed Method:
- Reduced Label setting explained with fixed and random parts.
- Classification risk formulation using reduced labels.
- Data augmentation and mixup training strategies implemented.
Effectiveness Analysis:
- Evaluation of annotation strength and browsing labels with Reduced Labels.
- Times of browsing labels compared between different settings.
Estimation Error Bound:
- Theoretical analysis of generalization estimation error bound for the proposed approach.
統計資料
Long-tailed distribution [4, 16, 23]
Extensive experiments conducted on benchmark datasets including ImageNet validate the effectiveness of our approach
引述
"Long-tailed data has received increasing attention in numerous real-world classification tasks."
"We introduce a novel weakly supervised labeling setting called Reduced Label."