The paper introduces Detector Collapse (DC), a groundbreaking backdoor attack paradigm tailored specifically for object detection (OD) tasks. Unlike previous OD backdoor attacks that primarily focused on localized errors, DC aims to indiscriminately degrade the overall performance of OD models.
To achieve this, the authors develop two innovative attack schemes:
SPONGE: This strategy triggers widespread misidentifications, flooding the output with a plethora of false positives. This overwhelms the computational resources of the detection system, leading to a significant reduction in processing speed and culminating in a denial-of-service.
BLINDING: This approach compromises the model's perception, causing it to classify all objects as the background, thereby rendering them 'invisible' to the OD system.
The paper also introduces a novel poisoning strategy that uses natural semantic features (e.g., a basketball) as triggers, enhancing the robustness of the backdoor in real-world environments. This is in contrast to previous works that relied on fixed-style triggers, which are less adaptable to dynamic real-world conditions.
Extensive evaluations on different detectors across several benchmarks demonstrate the significant improvement (up to 60% absolute and 7x relative) in attack efficacy of DC over state-of-the-art OD backdoor attacks. The authors also show that DC is resistant to potential defenses, such as fine-tuning and pruning.
Finally, the paper presents a physical-world demonstration of DC, showcasing its ability to catastrophically disable object detection systems in real-world settings.
他の言語に翻訳
原文コンテンツから
arxiv.org
深掘り質問