Core Concepts
Optimizing the instruction set for fine-tuning large language models by leveraging the complex interaction and dependency patterns between different categories of instructions.
Abstract
The paper investigates the interaction and dependency patterns between different categories of instructions, and proposes methods to optimize the instruction set and the fine-tuning process accordingly.
Key highlights:
- The authors systematically analyze the correlation patterns between different instruction categories using causal interventions. They find widespread positive and negative correlations, indicating the necessity of optimizing the category proportion beyond just selecting high-quality individual instructions.
- The authors induce an ability taxonomy of instructions based on causal interventions, revealing the hierarchical dependencies between different instruction categories. This suggests the need to arrange the learning order of instructions to ensure models acquire necessary preliminary skills before learning advanced ones.
- Leveraging the identified correlation patterns, the authors propose an effect equivalence-based linear programming method to optimize the category proportion of the instruction set. This outperforms strong baselines that only select high-quality individual instructions.
- Based on the induced ability taxonomy, the authors propose a curriculum learning approach that arranges the learning order of instruction categories. This further boosts the performance of language models compared to uniform mixing of instructions.
- Experiments on different language models demonstrate the generality and effectiveness of the proposed methods, supporting the reasonability of the induced interaction and dependency patterns.
Stats
"The equivalent total amount for an arbitrary category of instructions is |Ci| + Σj γji|Cj|, where |Ci| is the size of category i and γji is the effect equivalence coefficient measuring the correlation strength between category j and i."
"The proportion of each category of instructions is optimized by turning it into an effect-equivalence-based linear programming problem."
Quotes
"Emerging evidence (Dong et al.; Yuan et al., 2023) and our analyses indicate that complex correlation and dependency relationships exist between different categories of instructions. Therefore, considering the quality of individual instructions alone can be a suboptimal approach for building a fine-tuning instruction set."
"The dependency between skills necessitates that models acquire foundational knowledge before progressing to more complex tasks; otherwise, the effectiveness of instruction tuning will be compromised (Longpre et al., 2023)."