Yang, M., Gao, L., Li, P., Li, W., Dong, Y., & Cui, Z. (2024). Decay Pruning Method: Smooth Pruning With a Self-Rectifying Procedure. arXiv preprint arXiv:2406.03879v2.
This paper introduces a novel approach called the Decay Pruning Method (DPM) to address the limitations of traditional single-step pruning methods in compressing deep neural networks. The authors aim to improve the accuracy and efficiency of network pruning by mitigating the abrupt network changes and information loss associated with single-step pruning.
DPM consists of two key components: Smooth Pruning (SP) and Self-Rectifying (SR). SP replaces the abrupt single-step pruning with a gradual N-step process, gradually decaying the weights of redundant structures to zero while maintaining continuous optimization. SR leverages gradient information to identify and rectify sub-optimal pruning decisions during the SP process. The authors integrate DPM with three existing pruning frameworks: OTOv2, Depgraph, and Gate Decorator, and evaluate its performance on various models and datasets.
The integration of DPM consistently improves the performance of the tested pruning frameworks. DPM achieves higher accuracy than the original pruning methods while further reducing FLOPs in most scenarios. The authors demonstrate the effectiveness of DPM across various models (VGG16, VGG19, ResNet50, ResNet56), datasets (CIFAR10, CIFAR100, ImageNet), and pruning criteria.
DPM offers a more effective and adaptable approach to network pruning by combining gradual weight decay with a gradient-based self-rectifying mechanism. The method's generalizability and consistent performance improvements across different pruning frameworks highlight its potential as a valuable tool for compressing deep neural networks.
This research contributes to the field of model compression by introducing a novel pruning method that addresses the limitations of existing techniques. DPM's ability to improve both accuracy and efficiency has significant implications for deploying deep learning models on resource-constrained devices.
The paper primarily focuses on channel-wise pruning and evaluates DPM on image classification tasks. Further research could explore the effectiveness of DPM with other pruning granularities and applications beyond image classification. Additionally, investigating the optimal hyperparameter settings for DPM across different scenarios could further enhance its performance.
他の言語に翻訳
原文コンテンツから
arxiv.org
深掘り質問