MATEval: A Multi-Agent Framework for Reliable Evaluation of Open-Ended Text Generated by Large Language Models
The MATEval framework enhances the reliability and efficiency of evaluating open-ended text generated by large language models through a multi-agent discussion process that integrates self-reflection and Chain-of-Thought strategies, along with feedback mechanisms to reach consensus.