Alapfogalmak
SFTMix leverages training dynamics and a Mixup-based regularization technique to enhance the instruction-tuning process of large language models, leading to improved performance in instruction-following and domain-specific tasks without relying on perfectly curated datasets.
Statisztikák
SFTMix yields a greater improvement in multi-turn conversational abilities (with an average increase of 0.3 points) compared to single-turn performance (an average increase of 0.2 points) in MT-Bench.
In AlpacaEval-2, SFTMix shows a significant improvement in the length-controlled (LC) win rate.
SFTMix leads to a 1.33% absolute improvement (from 60.72% to 62.05%) for Llama-3.1-8B and a 1.66% increase (from 54.32% to 55.98%) for Mistral-7B-v0.1 in macro-average accuracy across four healthcare-related question-answering benchmarks.