The paper reveals the two-sided impact of data augmentation (DA) on closed-set and open-set recognition (OSR) performance. While multiple-sample-based augmentation (MSA) significantly boosts closed-set accuracy, it also leads to a substantial decline in OSR capability.
The authors investigate this phenomenon and find that MSA diminishes the criteria for OSR by reducing the magnitude of feature activations and logits, leading to greater uncertainty in distinguishing unknown samples. To mitigate this issue, the authors propose an asymmetric distillation framework that introduces extra raw data to the teacher model to enlarge its benefit on the mixed inputs. Additionally, a joint mutual information loss and a selective relabel strategy are utilized to encourage the student model to focus more on class-specific features within the mixed samples and decrease its activation on non-discriminative features.
Extensive experiments on various benchmarks demonstrate the effectiveness of the proposed method. It outperforms state-of-the-art open-set recognition methods by a significant margin while maintaining closed-set accuracy. The authors also show the generalization of their approach to other tasks like out-of-distribution detection and medical image analysis.
На другой язык
из исходного контента
arxiv.org
Дополнительные вопросы