The paper introduces LLMChain, a decentralized blockchain-based reputation system for sharing and evaluating Large Language Models (LLMs). LLMs have witnessed rapid growth in language understanding, generation, and reasoning capabilities, but they are susceptible to undesirable behaviors such as hallucinations, unreliable reasoning, and the generation of harmful content, which undermine trust in these models and pose challenges to their adoption in critical applications.
To address these issues, LLMChain combines automatic evaluation with human feedback to assign contextual reputation scores that accurately reflect the behavior of LLMs. The framework consists of four main layers:
The reputation model in LLMChain involves two stages: Interaction Evaluation and Global Scores Updating. The Interaction Evaluation stage computes an automatic score (Sa), a human score (Sh), and a weighted combination (Sθ) between both scores. The Global Scores Updating stage then updates the global reputation scores (Ra, Rh, and R) using predefined functions.
The experiments demonstrate the effectiveness of the automatic and human evaluation models, as well as the scalability and performance of the deployed blockchain network. LLMChain is the first decentralized framework for sharing and evaluating LLMs, and the authors have also released the LLMGooAQ dataset, a comprehensive dataset of 100k questions and answers generated by seven LLMs.
In un'altra lingua
dal contenuto originale
arxiv.org
Approfondimenti chiave tratti da
by Mouh... alle arxiv.org 04-23-2024
https://arxiv.org/pdf/2404.13236.pdfDomande più approfondite