This research paper presents a case study on using Retrieval-Augmented Generation (RAG) to improve the accuracy of answering domain-specific questions, focusing on Pittsburgh and Carnegie Mellon University (CMU).
Bibliographic Information: Sun, H., Wang, Y., & Zhang, S. (2024). Retrieval-Augmented Generation for Domain-Specific Question Answering: A Case Study on Pittsburgh and CMU. arXiv preprint arXiv:2411.13691v1.
Research Objective: The study aimed to design and implement a RAG system that enhances the ability of large language models to answer questions about Pittsburgh and CMU accurately.
Methodology: The researchers collected data from various websites related to Pittsburgh and CMU, employing a greedy scraping strategy. They annotated the data using a combination of manual annotation and automatic generation via the Mistral model. The RAG framework integrated BM25 and FAISS retrievers, enhanced with a reranker for improved document retrieval accuracy. The researchers used a 7B Mistral model as the backbone LLM and evaluated the system's performance using metrics like Exact Match (EM), Precision, Recall, and F1 Score.
Key Findings: The RAG system significantly outperformed a non-RAG baseline, particularly in answering time-sensitive and complex queries. The best RAG configuration, which included a document re-ranker, few-shot learning, and an ensembled retriever, achieved an F1 score of 42.21%, a substantial improvement over the baseline's 5.45%.
Main Conclusions: The study demonstrates the potential of RAG systems in enhancing answer precision and relevance for domain-specific question answering. The researchers highlight the importance of document retrieval accuracy and the effectiveness of few-shot learning in improving the system's performance.
Significance: This research contributes to the growing field of RAG and its application in domain-specific question answering. The findings have implications for developing more accurate and efficient question-answering systems, particularly for domains requiring up-to-date information.
Limitations and Future Research: The study acknowledges limitations regarding the dataset's generalizability and the potential for errors due to incorrect document retrieval. Future research could focus on expanding the dataset, refining retrieval methods, and exploring the impact of different LLM architectures on the system's performance.
翻译成其他语言
从原文生成
arxiv.org
更深入的查询