Alapfogalmak
Instructing large language models to identify and ignore irrelevant conditions improves accuracy in solving math word problems.
Kivonat
This article introduces a novel approach, I3C, that instructs large language models (LLMs) to identify and ignore irrelevant conditions in math word problems. By selecting the most confusing problems as demonstrations, I3C significantly enhances the performance of LLMs in solving complex multi-step reasoning tasks. Extensive experiments on eight math word problem datasets demonstrate the effectiveness and efficiency of this method.
Abstract:
- Existing chain-of-thought prompting methods struggle with irrelevant conditions.
- I3C proposes a novel approach to instruct LLMs to identify and ignore irrelevant conditions.
- Demonstrations are used to enhance few-shot learning abilities.
Introduction:
- Math word problem solving requires multi-step reasoning abilities.
- CoT prompting methods guide LLMs but can be confused by irrelevant conditions.
- I3C aims to improve reasoning paths by identifying and ignoring irrelevant conditions.
Proposed Approach:
- I3C identifies irrelevant condition candidates based on semantic relevance.
- LLMs verify if candidates are indeed irrelevant.
- A novel instruction is created to help LLMs avoid confusion and improve reasoning paths.
Experiments:
- I3C combined with CoT methods improves performance on various MWP datasets.
- I3C-Select outperforms existing prompting methods by selecting confusing problems as demonstrations.
Statisztikák
I3C achieves an accuracy of 96.0% on GSM-IC2-1K with GPT-3.5-Turbo.
I3C significantly outperforms Complex-CoT by +11.7% on GSM-IC2-1K.
Idézetek
"Instructing Large Language Models to Identify and Ignore Irrelevant Conditions."
"I3C can be combined with any CoT prompting methods to improve the performance of solving MWPs."