ConstitutionalExperts presents a novel approach to prompt optimization using constitutional principles, incrementally refining prompts for better performance. By training unique prompts for different semantic regions and employing a mixture-of-experts architecture, the method achieves superior results compared to existing techniques across various benchmark datasets.
Large language models (LLMs) excel at NLP tasks with appropriate prompts but face challenges in prompt creation. ConstitutionalExperts introduces a method that incrementally refines prompts based on constitutional principles, leading to improved performance. The technique involves clustering training data, training unique experts for each cluster, and routing inputs at inference time using similarity measures.
The method is evaluated across six benchmark datasets and shows significant improvement over state-of-the-art prompt optimization techniques. By structuring prompts as lists of principles and training unique experts for different semantic regions, ConstitutionalExperts achieves better performance and interpretability. The inclusion of a mixture-of-experts architecture further enhances the overall effectiveness of the approach.
Results indicate that ConstitutionalExperts outperforms existing methods by 10.9% in F1 score and demonstrates the broad applicability of the mixture-of-experts architecture in improving prompt optimization techniques. Future work could explore additional NLP tasks, alternative clustering methods, and human interventions to guide expert edits.
다른 언어로
소스 콘텐츠 기반
arxiv.org
더 깊은 질문