SLTrain: Enhancing Large Language Model Pretraining with Sparse and Low-Rank Techniques for Improved Memory and Parameter Efficiency
SLTrain introduces a novel approach to pretraining large language models (LLMs) by combining sparse and low-rank matrix factorization, achieving comparable performance to full-rank training while significantly reducing memory and parameter requirements.