Core Concepts
NEPENTHE, an iterative unstructured approach, can effectively reduce the depth of over-parameterized deep neural networks with little to no performance loss.
Abstract
The paper presents NEPENTHE, a method that aims to reduce the depth of over-parameterized deep neural networks. The key insights are:
The authors show that unstructured pruning naturally minimizes the entropy of rectifier-activated neurons, which can be used to identify layers that can be removed entirely.
They propose an entropy-weighted pruning score that guides the pruning process to favor the removal of layers with low entropy.
NEPENTHE iteratively prunes the network, removing layers with zero entropy without significant performance degradation.
The authors validate NEPENTHE on popular architectures like MobileNet and Swin-T across various datasets. They demonstrate that NEPENTHE can effectively linearize some layers, reducing the model's depth, while maintaining high performance, especially when dealing with over-parameterized networks.
Stats
The paper does not provide any specific numerical data or metrics in the main text. The results are presented in a tabular format.
Quotes
The paper does not contain any striking quotes that support the key logics.