핵심 개념
The author explores generating math word problems that challenge large language models, aiming to degrade their problem-solving abilities significantly while maintaining the original difficulty and coherence of the questions.
초록
The content discusses a novel approach to creating math word problems that are unsolvable by large language models (LLMs) to ensure fair evaluation in education. By leveraging abstract syntax trees, the method generates adversarial examples that cause LLMs to produce incorrect answers by editing numeric values in the problems. The study evaluates various LLMs, proposing a cost-effective approach to attacking high-cost models and conducting human evaluations on the generated adversarial examples.
Key points include:
- Introduction of a new paradigm for fair evaluation in education.
- Use of abstract syntax trees to generate unsolvable math word problems for LLMs.
- Evaluation of different LLMs' performance under adversarial settings.
- Proposal of a cost-effective method for attacking expensive API-based models.
- Human evaluation results indicating correctness, coherence, and similarity of generated problems.
The study aims to shed light on educational tool development and ethical use of LLMs in education.
통계
We conduct experiments on 7 open-source models: MetaMath 7B, Mistral 7B, Llama-2 13B, WizardMath 13B, Vicuna 13B, CodeLlama 34B, MetaMath 70B.
We also evaluate 2 closed-source models: GPT-4-Turbo and GPT-3.5-Turbo.
인용구
"Generating adversarial examples which preserve the structure and difficulty of the original questions aimed for assessment."
"Focusing on math word problems to structurally generate adversarial examples causing LLMs to produce incorrect answers."