The paper introduces a Feature Corrective Transfer Learning (FCTL) framework to address the challenge of robust object detection under non-ideal visual conditions, such as rain, fog, low illumination, or raw Bayer images without ISP processing.
The key aspects of the methodology are:
Initial training of a comprehensive object detection model (Faster R-CNN) on a pristine RGB dataset to establish a strong baseline.
Generation of non-ideal image versions (e.g., rainy, foggy, low-light, raw Bayer) from the original dataset.
Fine-tuning of the same model on the non-ideal images, but with a novel loss function called Extended Area Novel Structural Discrepancy Loss (EANSDL) that compares the feature maps of the model trained on ideal and non-ideal images. This allows for direct feature map correction without modifying the underlying model architecture.
The EANSDL loss function adaptively balances the analysis between detailed pixel-level discrepancies and broader spatial pattern alignments, dynamically adjusting the gradient consistency evaluation across the feature pyramid's hierarchical layers.
The proposed Non-Ideal Image Transfer Faster R-CNN (NITF-RCNN) model, which incorporates the FCTL approach, demonstrates significant improvements in mean Average Precision (mAP) compared to the baseline Faster R-CNN model, with relative gains of 3.8-8.1% under various non-ideal conditions. The model's performance on non-ideal datasets also approaches that of the baseline on the original ideal dataset, showcasing its robustness and versatility.
To Another Language
from source content
arxiv.org
Key Insights Distilled From
by Chuheng Wei,... at arxiv.org 04-18-2024
https://arxiv.org/pdf/2404.11214.pdfDeeper Inquiries