NOLA, a novel reparameterization technique, enables efficient fine-tuning of large language models by decoupling the number of trainable parameters from the model architecture and the chosen rank, breaking the rank-one decomposition limit of existing methods like LoRA.
HydraLoRA, an asymmetric LoRA architecture, enhances the efficiency and performance of fine-tuning large language models by leveraging a shared matrix A and multiple distinct matrices B to capture both common and task-specific knowledge.
Birbal, a Mistral-7B based model, achieved a 35% performance improvement through high-quality instruction curation.