The paper presents the ASC2End system, a novel tool for automating large-scale information comparison. Key highlights:
ASC2End was developed to address the challenges of applying generative language models (LLMs) for information comparison at scale, such as maintaining information across large contexts and overcoming model token limitations.
The system comprises four modules: Document Summarization (DS), Criteria Embedding (CE), Retrieval Augmented Generation (RAG), and Comparison Assessment (CA).
The DS module performs abstractive summarization on individual documents to generate concise summaries. The CE module embeds and splits the user-defined criteria document for efficient retrieval.
The RAG module uses semantic text similarity to retrieve the most relevant criteria passages for the document summary. The CA module then compares the summary to the retrieved criteria and generates a detailed analysis.
Experiments were conducted using financial news data and sustainability criteria. Machine-level LLMs were evaluated for the summarization task using ROUGE scores, while human-level LLMs were assessed for the comparison assessment through a user survey.
The results demonstrate the effectiveness of the ASC2End system in automating large-scale information comparison, providing time-saving efficiencies and enabling more informed decision-making.
Til et annet språk
fra kildeinnhold
arxiv.org
Viktige innsikter hentet fra
by Truman Yuen,... klokken arxiv.org 04-09-2024
https://arxiv.org/pdf/2404.04351.pdfDypere Spørsmål