Detecting and Mitigating Hallucinations in Large Language Model-Generated Content: Insights from Human Perception and Engagement
Humans can discern the relative accuracy of LLM-generated content, ranking it as genuine > minor hallucination > major hallucination. Warning labels reduce the perceived accuracy and increase dislike of hallucinated content, without significantly affecting genuine content.