핵심 개념
LoRITa promotes low-rankness in neural network weights through the composition of linear layers during training, enabling efficient post-training compression without changing the network structure.
초록
The paper introduces a novel approach called Low-Rank Induced Training (LoRITa) that promotes low-rankness in neural network weights during training. LoRITa achieves this by overparameterizing the layers to be compressed through linear layer composition, followed by post-training singular value truncation on the product of the composed weights.
The key highlights are:
- LoRITa eliminates the need for initializing with pre-trained models or specifying the rank prior to training, unlike previous low-rank training methods.
- Theoretical justification is provided, showing that standard weight decay regularization naturally imposes low-rankness on models with linear layer composition before activation.
- Extensive experiments on image classification tasks using MNIST, CIFAR10, and CIFAR100 datasets demonstrate the effectiveness of LoRITa across different neural network architectures, including Fully Connected Networks (FCNs), Convolutional Neural Networks (CNNs), and Vision Transformers (ViTs).
- Compared to leading structured pruning methods, LoRITa achieves either competitive or state-of-the-art results in terms of FLOPs reduction and parameter drop.
통계
The paper does not provide specific numerical data points to support the key logics. The results are presented in the form of plots and comparative tables.
인용구
The paper does not contain any striking quotes that support the key logics.