Core Concepts
LLMChain is a decentralized blockchain-based reputation system that combines automatic evaluation with human feedback to assign contextual reputation scores to accurately reflect the behavior of Large Language Models (LLMs), enabling users to identify the most trustworthy LLM for their needs and providing LLM developers with valuable insights to refine and improve their models.
Abstract
The paper introduces LLMChain, a decentralized blockchain-based reputation system for sharing and evaluating Large Language Models (LLMs). LLMs have witnessed rapid growth in language understanding, generation, and reasoning capabilities, but they are susceptible to undesirable behaviors such as hallucinations, unreliable reasoning, and the generation of harmful content, which undermine trust in these models and pose challenges to their adoption in critical applications.
To address these issues, LLMChain combines automatic evaluation with human feedback to assign contextual reputation scores that accurately reflect the behavior of LLMs. The framework consists of four main layers:
- User Layer: Individuals with different areas of expertise can use shared LLMs and provide feedback on their interactions.
- Blockchain Layer: A permissioned blockchain network managed by LLM providers and developers, where LLMs are shared and evaluated.
- Oracle Layer: A decentralized network that automates the evaluation process, intercepting responses from models, conducting off-chain automatic evaluations, and triggering on-chain smart contracts to update the overall score of the targeted LLM.
- LLM Layer: Language models administered locally by LLM providers and developers.
The reputation model in LLMChain involves two stages: Interaction Evaluation and Global Scores Updating. The Interaction Evaluation stage computes an automatic score (Sa), a human score (Sh), and a weighted combination (Sθ) between both scores. The Global Scores Updating stage then updates the global reputation scores (Ra, Rh, and R) using predefined functions.
The experiments demonstrate the effectiveness of the automatic and human evaluation models, as well as the scalability and performance of the deployed blockchain network. LLMChain is the first decentralized framework for sharing and evaluating LLMs, and the authors have also released the LLMGooAQ dataset, a comprehensive dataset of 100k questions and answers generated by seven LLMs.
Stats
The paper does not provide specific numerical data or statistics. It focuses on the design and implementation of the LLMChain framework.
Quotes
"LLMs often inherit biases present in their training data, reflecting societal prejudices and stereotypes [6]. Consequently, these models can produce outputs that perpetuate or even exacerbate existing social inequalities."
"LLMs may also display unreliable reasoning [9], characterized by a lack of consistent or dependable logical abilities."
"These flawed actions that diminish trust in LLMs cause users to be cautious about relying on AI-generated content due to its unpredictability and potential for producing incorrect information."