Core Concepts
QDA-SQL is a novel data augmentation method that improves the performance of Large Language Models (LLMs) on multi-turn Text-to-SQL tasks by generating diverse and challenging training samples, including ambiguous and unanswerable questions.
Stats
The researchers generated 10,874 dialogues, consisting of 65,393 turns, using the QDA-SQL framework.
The SQL generated in the enhanced dataset demonstrated greater tree depths compared to the original dataset, indicating the generation of more complex queries.
Manual review showed that Gemini Pro accurately identified 94% of the misclassified samples during the "Verify and Refine" process.
Evaluation using GPT-4 indicated that 62% of the QDA-SQL enhanced dataset was considered superior to the manually annotated original SParC and CoSQL training sets.
Models fine-tuned with the augmented dataset showed significant improvements in QM, especially for questions with higher difficulty levels.
The enhanced training dataset led to marked improvements in QM for multi-turn dialogues across various difficulty levels.
Models trained with QDA-SQL equipped with Intent Recognition samples achieved an average F1 score of 57.8 on intent recognition tasks.
Removing the Verify state in the StateFlow model resulted in lower AccS and IAccS values and higher error rates.
Removing the Intent Recognition state led to reduced IAccS, particularly for questions not answerable using SQL.
Error analysis revealed that the fine-tuned CodeLlama model demonstrated the lowest error rate, with a marked reduction in logical and database comprehension errors compared to GPT-4 and the unfine-tuned CodeLlama.
Quotes
"Existing approaches predominantly focus on enhancing the SQL generation ability of LLMs, without consideration of multiple types of questions. This may result in incorrect responses for questions that cannot be answered using SQL."
"Our research endeavors to leverage LLM to refine data augmentation for Text-to-SQL by generating a more diverse and natural set of training samples, thereby enhancing their practical applicability."
"By formatting the dataset to align with the states in the StateFlow model, we ensure logical coherence and effective training."