Liu, T., Hou, L., Wang, L., Song, X., & Yan, B. (2025). Dense Optimizer: An Information Entropy-Guided Structural Search Method for Dense-like Neural Network Design. Journal of LaTeX Class Files, 14(8).
This paper introduces Dense Optimizer, a novel method for automatically designing efficient dense-like neural network architectures by formulating the process as an optimization problem guided by information entropy and power-law distribution principles.
The researchers define the structural entropy of a DenseBlock, considering information reuse and concatenation operations. They introduce an effectiveness metric based on network depth and width. Observing that the information entropy distribution in dense networks follows a power-law, they incorporate this principle as a constraint in the optimization model. A branch-and-bound algorithm is proposed to efficiently search for the optimal network configuration by maximizing information entropy while adhering to the power-law distribution and computational constraints. The optimized architectures are then trained and evaluated on CIFAR-10, CIFAR-100, and SVHN datasets.
Dense Optimizer successfully designs dense-like networks that outperform manually designed counterparts and other NAS methods in terms of accuracy under various computational budgets. The optimized models demonstrate significant improvements over the original DenseNet on benchmark image classification tasks. Ablation studies confirm the importance of the power-law constraint in achieving superior performance.
Dense Optimizer offers an efficient and effective alternative to manual design and computationally expensive NAS methods for dense-like neural networks. The proposed method leverages information entropy and power-law distribution principles to guide the search process, resulting in high-performing architectures.
This research contributes to the field of Neural Architecture Search by introducing a novel optimization-based approach specifically tailored for dense-like networks. The use of information entropy and power-law distribution as guiding principles provides valuable insights for designing efficient and effective deep learning models.
The study primarily focuses on image classification tasks and traditional dense-BC convolutional blocks. Future research could explore the applicability of Dense Optimizer to other tasks, network architectures, and convolutional block designs. Investigating the generalization capabilities of the proposed method across different datasets and domains would be beneficial.
翻譯成其他語言
從原文內容
arxiv.org
深入探究