The paper introduces a pioneering method called "vocabulary-defined semantics" to analyze the semantics of language model (LM) latent space. The key highlights are:
Semantic Basis: The authors define "semantic basis" by obtaining the representations of vocabulary labels using the LM head matrix pseudoinverse. This establishes a disentangled reference frame within the LM latent space.
Semantic Feature: The authors propose a novel "Vocabulary Affinity Inference" (VAI) method to compute logits based on distance-based similarities with the semantic bases, leveraging the differentiability and local isotropy of transformer models.
Semantic Calibration: The authors regard LM adaptation as a process of calibrating the semantics of data representations. They introduce a lightweight neural clustering module to refine the representations by clustering them around the semantic bases.
The authors conduct extensive experiments across diverse text understanding datasets and LLM scales, demonstrating that their approach outperforms state-of-the-art methods in retrieval-augmented generation and parameter-efficient finetuning. The findings not only shed light on LM mechanics but also offer practical solutions to enhance LM performance and interpretability.
Para outro idioma
do conteúdo fonte
arxiv.org
Principais Insights Extraídos De
by Jian Gu,Alde... às arxiv.org 04-09-2024
https://arxiv.org/pdf/2401.16184.pdfPerguntas Mais Profundas