Kernekoncepter
Concept-aware data construction improves in-context learning by training language models to utilize latent concepts from demonstrations.
Resumé
Concept-aware Training (CoAT) is proposed to enhance in-context learning by focusing on concept-dependent training data. CoAT improves the utilization of new latent concepts and enhances robustness to functional deficiencies. The framework challenges language models to learn and apply latent reasoning concepts from demonstrations, leading to improved performance on diverse tasks. CoAT outperforms traditional instruction tuning approaches and achieves comparable results with significantly less training data. The method shows promise for democratizing the creation of accurate in-context learners for various applications.
Statistik
Recent LMs are based on 10 to 100 times smaller models but achieve comparable ICL quality.
CoAT enables ICL of otherwise not learnable tasks with only two training tasks.
CoAT-trained models show significant improvements over traditional instruction tuning approaches.
CoAT models reach higher accuracy on a majority of tasks compared to baselines.
CoAT outperforms multitask learners on a majority of tasks in various evaluations.
Citater
"Many recent language models are capable of in-context learning, manifested in the LMs’ ability to perform a new task solely from a natural-language instruction."
"We propose Concept-aware Training (CoAT), a framework for constructing training scenarios that make it beneficial for the LM to learn to utilize the analogical reasoning concepts from demonstrations."
"Finally, we show that concept-aware in-context learning is more effective for a majority of new tasks when compared to traditional instruction tuning."