Automatic Evaluation of Human-Model Interactive Question Answering Using LLM-Based Evaluation Agents (IQA-EVAL)
This research paper introduces IQA-EVAL, a novel framework for automatically evaluating interactive question answering (IQA) systems using Large Language Model (LLM)-based Evaluation Agents (LEAs) that simulate human interaction and judgment, offering a cost-effective and scalable alternative to traditional human evaluation methods.