This research paper presents an empirical evaluation of response uncertainty in Multimodal Large Language Models (MLLMs) when presented with misleading information. The authors argue that existing benchmarks for evaluating MLLMs primarily focus on response correctness and overlook the uncertainty inherent in these responses, especially when misleading information is introduced.
To address this gap, the authors propose a two-stage pipeline to identify and quantify response uncertainty. First, they collect MLLM responses to image-question pairs without misleading information. Then, they introduce misleading instructions, either explicitly stating the wrong answer or implicitly suggesting it, and observe the changes in the model's responses. This approach allows them to identify data points where the model exhibits uncertainty and quantify this uncertainty using a metric called the "misleading rate."
Based on their findings, the authors construct a new benchmark called the Multimodal Uncertainty Benchmark (MUB), specifically designed to evaluate MLLM robustness against misleading information. They categorize the benchmark data into three difficulty levels based on the number of models misled by the misleading instructions.
The authors evaluate 12 open-source and 5 close-source MLLMs on MUB and find that all models are highly susceptible to both explicit and implicit misleading instructions. They observe that larger models are not necessarily more robust and that even models with high confidence scores can be easily misled.
To mitigate this vulnerability, the authors propose a fine-tuning strategy using a mix of explicit and implicit misleading instructions. Their experiments demonstrate that this strategy significantly reduces the misleading rate across all tested models without compromising their performance on other tasks.
The paper concludes by emphasizing the importance of incorporating misleading information during MLLM training to enhance their robustness and reliability for real-world applications. The authors also highlight the need for further research into understanding and mitigating the impact of misleading information on MLLM performance.
To Another Language
from source content
arxiv.org
Key Insights Distilled From
by Yunkai Dang,... at arxiv.org 11-06-2024
https://arxiv.org/pdf/2411.02708.pdfDeeper Inquiries