AC-EVAL evaluates Large Language Models' proficiency in ancient Chinese language understanding, highlighting areas for improvement.
다국어 기반 모델의 언어, 문화, 추론 능력 평가
言語モデルの文化的知識を評価するためのHAE-RAE Benchの導入
Challenging non-Korean language models with cultural and contextual knowledge through the HAE-RAE Bench dataset.
LLMs show potential for enhancing ancient text comprehension, as highlighted by AC-EVAL.
AC-EVAL aims to assess the proficiency of Large Language Models in understanding ancient Chinese through a comprehensive benchmark, highlighting areas for improvement and development.