Core Concepts
Mixtures of In-Context Learners (MOICL) is a novel approach that enhances the effectiveness and efficiency of in-context learning (ICL) in large language models by treating subsets of demonstrations as experts and dynamically learning their optimal weighting during training, leading to improved performance, robustness, and reduced computational complexity.
Stats
MOICL achieves up to 13% accuracy improvement compared to standard ICL and LENS on 5 out of 7 classification datasets.
When 70% of the demonstrations are out-of-domain, MOICL shows a minimal performance drop compared to standard ICL.
In the presence of label imbalance, MOICL experiences a significantly smaller performance drop compared to standard ICL.
With 10 noisy demonstrations, MOICL maintains performance while standard ICL experiences a significant drop.
MOICL achieves better performance than standard ICL with only around 20 annotated demonstrations, highlighting its data efficiency.
MOICL consistently shows higher accuracy relative to inference time compared to standard ICL, demonstrating its time efficiency.