Khái niệm cốt lõi
The core message of this paper is to propose a novel model called 'Logic-Query-of-Thoughts' (LGOT) that combines the strengths of Large Language Models (LLMs) and knowledge graph reasoning to effectively answer complex logic queries.
Tóm tắt
The paper addresses the limitations of both Large Language Models (LLMs) and Knowledge Graph Question Answering (KGQA) methods when it comes to answering complex logic queries. LLMs tend to suffer from the hallucination problem and struggle with factual recall, while KGQA methods deteriorate quickly when the underlying knowledge graph is incomplete.
To overcome these challenges, the authors introduce the 'Logic-Query-of-Thoughts' (LGOT) framework. LGOT seamlessly integrates knowledge graph reasoning and LLMs, breaking down complex logic queries into easier-to-answer subquestions. It utilizes both knowledge graph reasoning and LLMs to derive answers for each subquestion, and then aggregates the results to select the highest quality candidate answers.
The key components of LGOT include:
- Interfaces for LLMs and KGQA to perform logical operations like projection and intersection.
- A method to combine the outputs of LLMs and KGQA, leveraging the likelihood ratio test and fuzzy vector representations.
- Techniques to guide LLMs in accordance with the logic query structure, including relation parsing and prompt engineering for projection and intersection operations.
- An optional answer evaluation module that employs LLMs to assess the quality of the generated responses.
The experimental results on various real-world datasets demonstrate that LGOT significantly outperforms state-of-the-art baselines, including ChatGPT, Chain-of-Thought, and knowledge graph reasoning methods, with up to 20% improvement in performance.
Thống kê
"The knowledge graph for the MetaQA dataset contains 43,234 entities and 18 relations, with 150,160 training edges and 8,000 test edges. The 50% incomplete knowledge graph has 66,791 training edges and 4,000 test edges."
"The knowledge graph for the ComplexWebQuestions dataset contains 81,272 entities and 338 relations, with 423,663 training edges and 35,558 test edges. The 50% incomplete knowledge graph has 245,876 training edges and 35,558 test edges."
"The knowledge graph for the GraphQuestions dataset contains 64,625 entities and 715 relations, with 70,291 training edges and 14,015 test edges. The 50% incomplete knowledge graph has 35,145 training edges and 14,059 test edges."
Trích dẫn
"LLMs tend to memorize facts and knowledge present in their training data (Petroni et al., 2019). However, research has revealed that LLMs struggle with factual recall and could generate factually incorrect statements, leading to hallucinations (Pan et al., 2023)."
"Different from Large Language Models (LLMs), knowledge graphs store structured human knowledge, making them a valuable resource for finding answers. Knowledge Graph Question Answering (KGQA) (Liu et al., 2023b, 2022a, 2023a) aims to identify an answer entity within the knowledge graph to respond to a given question. Compared with LLMs, KGQA generates more accurate results when the knowledge graph is complete. However, the performance of KGQA deteriorates quickly when the underlying KG itself is incomplete with missing relations."