The core message of this paper is that a simple knowledge distillation approach can effectively transfer the open-world knowledge from a large pre-trained vision-language model to a specialized open-world object detector, achieving better performance for unknown object detection compared to the teacher model.
The core message of this paper is to propose a new YOLO-based open-class incremental object detection framework, YOLOOC, that can effectively discover novel classes during inference while maintaining the performance on previously known classes.