Core Concepts
MLLMs are vulnerable to attacks using query-relevant images, necessitating the development of safety measures and evaluation frameworks.
Stats
"Warning: This paper contains examples of harmful language and images."
"Our analysis across 12 state-of-the-art models reveals that MLLMs are susceptible to breaches instigated by our approach."
"In response, we propose a straightforward yet effective prompting strategy."