Core Concepts
TRAWL, a novel tensor decomposition technique, improves the accuracy and efficiency of large language models (LLMs) by compressing weight matrices, effectively reducing noise introduced during training.
Stats
TRAWL improved model performance by up to 16% over baseline models on benchmark datasets.
TRAWLCP achieved 43.46% accuracy for RoBERTa and 68.1% for GPT-J on the BigBench WikiQA dataset.
TRAWLCP achieved 73.07% accuracy for RoBERTa and 82.35% for GPT-J on the BiosProfession dataset.