ImgTrojan: Exploiting VLM Vulnerabilities with ONE Image
The author introduces ImgTrojan, a novel jailbreaking attack against Vision-Language Models (VLMs) by poisoning training data with malicious image-text pairs. This method effectively bypasses safety barriers and highlights the vulnerability of VLMs.