The authors propose a method called EASIER (Entropy-bASed Importance mEtRic) to reduce the depth of over-parameterized deep neural networks. The key idea is to identify layers that are close to becoming linear by estimating the entropy of the rectifier activations in each layer. The layer with the lowest entropy is then replaced with a linear activation, effectively reducing the depth of the network.
The method works as follows:
The authors evaluate EASIER on four popular models (ResNet-18, MobileNetv2, Swin-T, and VGG-16) across seven datasets for image classification. They compare the results to two existing methods: Layer Folding and EGP (an entropy-guided pruning technique).
The results show that EASIER can consistently produce models with better performance for the same number of layers removed, compared to the other methods. It is also able to remove more layers while maintaining similar performance to the original model. The authors also provide an ablation study on the choice of rectifier activation and the feasibility of a one-shot approach.
Para Outro Idioma
do conteúdo original
arxiv.org
Principais Insights Extraídos De
by Vict... às arxiv.org 05-01-2024
https://arxiv.org/pdf/2404.18949.pdfPerguntas Mais Profundas