This paper proposes an active learning framework to effectively and efficiently mitigate hallucinations in large language models (LLMs) for text summarization by selecting diverse hallucination samples for annotation and finetuning.
Effective strategies for mitigating hallucinations in large language models through targeted interventions in specific model components.
Large language models can confidently generate responses that are incorrect or nonsensical (hallucinations). This work proposes a principled procedure to determine when the model should abstain from responding, instead of hallucinating, by leveraging the model's self-consistency as a measure of confidence.
Rowen, a novel framework that enhances large language models (LLMs) with an adaptive retrieval augmentation process to effectively mitigate hallucinated outputs.