核心概念
Proposing a three-phases supervised fine-tuned model with a strong prior module for educational knowledge disassembly and incremental guided output.
摘要
The content introduces a novel three-phases supervised fine-tuned (SFT) model for education, emphasizing the importance of a strong prior module. The model aims to provide step-by-step guidance to students by breaking down educational knowledge logically. It incorporates data classification, overlap estimation, pre-trained models, and a prior module to enhance tutoring capabilities. Extensive experiments demonstrate the model's state-of-the-art performance in coding abilities and conversational skills.
Structure:
- Abstract & Introduction:
- Proposal of an end-to-end SFT educational model.
- Evolution of language models from Markov chain to Transformer architecture.
- Methodology:
- Data preprocessing using an overlap estimation network.
- Three-phases LORA fine-tuning process.
- Structured FCN cutting and regularization.
- Implementation of the prior module for enhanced inference.
- Experimental Results:
- Evaluation of coding, chat, and tutoring abilities on various benchmarks.
- Comparison & Ablation Test:
- Comparative analysis of different model architectures.
- Ablation experiments to assess the impact of specific modules on performance.
統計資料
Extensive experiments report that our model achieves 75.10% accuracy on the HumanEval benchmark.
Our model maintains strong conversational capabilities with scores of 56.34, 50.60, and 45.27 on MMLU, C-Eval, and AGIEval benchmarks respectively.
引述
"Our model represents the first research effort to truly embody the tutor role with abundant educational knowledge."
"Extensive experiments demonstrate our model's state-of-the-art performance in coding abilities compared to open-source models."