Core Concepts
CRE-LLM, a framework for domain-specific Chinese relation extraction, leverages fine-tuned open-source large language models to directly and efficiently extract relations between entities in unstructured Chinese text.
Abstract
The paper introduces CRE-LLM, a novel framework for domain-specific Chinese relation extraction (DSCRE) that utilizes fine-tuned open-source large language models (LLMs) to directly extract relations between entities in unstructured Chinese text.
The key highlights are:
CRE-LLM addresses the challenges of complex network design, poor internal perception, and high fine-tuning costs faced by previous DSCRE methods. It employs a simple and direct generative approach by fine-tuning open-source LLMs like Llama-2, ChatGLM2, and Baichuan2.
The framework constructs an appropriate prompt based on the given entities and text, and then fine-tunes the LLMs using the Parameter-Efficient Fine-Tuning (PEFT) framework. This enhances the logical awareness and generation capabilities of the model for DSCRE tasks.
Extensive experiments on two domain-specific DSCRE datasets, FinRE and SanWen, demonstrate that CRE-LLM outperforms existing methods and achieves state-of-the-art performance on the FinRE dataset.
The fine-tuning approach using PEFT significantly reduces the memory consumption and environmental configuration requirements compared to full fine-tuning, making CRE-LLM more accessible for general projects and teams.
Error analysis reveals that the primary challenges lie in understanding entity relations, handling multiple relations, and correctly identifying the "NA" relation, which the authors aim to address in future work.
Overall, CRE-LLM represents a promising direction for applying powerful LLMs to domain-specific relation extraction tasks in a simple, efficient, and effective manner.
Stats
"With the establishment of [Ant Financial], [Alibaba]'s layout in the financial business has been officially clarified."
"内部人士昨日透露,[双汇国际]内部对于"双币双股"这种模式上市还没有完全确定。"
Quotes
"CRE-LLM enhances the logic-awareness and generative capabilities of the model by constructing an appropriate prompt and utilizing open-source LLMs for instruction-supervised fine-tuning."
"The experimental results show that CRE-LLM is significantly superior and robust, achieving state-of-the-art (SOTA) performance on the FinRE dataset."