Kun is a novel self-training method that leverages instruction back-translation and answer polishment to automatically generate large-scale, high-quality Chinese instruction-tuning datasets for LLMs, reducing the reliance on manual annotation.
This paper introduces IDEA-MCTS, a novel framework that leverages Monte Carlo Tree Search (MCTS) and evaluation models to automatically generate high-quality, diverse, and complex instruction data for improving the instruction-following abilities of large language models (LLMs).