Soylu, D., Potts, C., & Khattab, O. (2024). Fine-Tuning and Prompt Optimization: Two Great Steps that Work Better Together. arXiv preprint arXiv:2407.10930v2.
This paper investigates optimizing both language model weights and prompt templates in modular NLP pipelines to maximize downstream task performance, addressing the challenge of limited labeled data and computational resources.
The researchers propose the "BetterTogether" algorithm, which alternates between fine-tuning LM weights and optimizing prompt templates using bootstrapping strategies. They evaluate this approach on three NLP tasks: multi-hop question answering (HotPotQA), mathematical reasoning (GSM8K), and feature-based classification (Iris), using three different language models (Mistral, LLaMa-2, LLaMa-3).
The BetterTogether strategies, which combine prompt and weight optimization, consistently outperform strategies that optimize only prompts or weights. This approach leads to performance improvements of 5-78% on HotPotQA, 2.5-10% on GSM8K, and 3.5-88% on Iris, compared to single optimization techniques.
The study demonstrates the effectiveness of alternating prompt optimization and fine-tuning for improving the performance of modular language model pipelines. This approach enables language models to "teach themselves" and achieve better results than optimizing either component in isolation.
This research contributes to the growing field of optimizing complex language model pipelines, offering a practical and effective strategy for enhancing performance on diverse NLP tasks.
The study primarily focuses on LoRA fine-tuning and a limited set of tasks and language models. Future research could explore other fine-tuning methods and evaluate the generalizability of the BetterTogether approach across a wider range of NLP tasks and models. Further investigation is needed to understand the underlying mechanisms driving the synergy between prompt optimization and fine-tuning.
Іншою мовою
із вихідного контенту
arxiv.org
Ключові висновки, отримані з
by Dilara Soylu... о arxiv.org 10-08-2024
https://arxiv.org/pdf/2407.10930.pdfГлибші Запити