核心概念
SortedNet proposes a generalized solution for training many-in-one neural networks, leveraging modularity and stochastic updating to achieve efficient and scalable training.
要約
SortedNet introduces a novel approach to training dynamic neural networks by sorting sub-models based on computation/accuracy requirements. This method enables efficient switching between sub-models during inference, leading to superior performance over existing dynamic training methods. Extensive experiments across various architectures and tasks demonstrate the effectiveness and scalability of SortedNet in preserving model performance while reducing storage requirements and enabling dynamic inference capabilities.
統計
SortedNet is able to train up to 160 sub-models at once, achieving at least 96% of the original model’s performance.
SortedNet outperforms state-of-the-art methods in dynamic training on CIFAR10.
SortedNet offers minimal storage requirements and dynamic inference capability during inference.
引用
"For every minute spent organizing, an hour is earned." - Benjamin Franklin.
"SortedNet enables the training of numerous sub-models simultaneously, simplifies dynamic model selection and deployment during inference, and reduces the model storage requirement significantly."
"Our method outperforms previous dynamic training methods and yields more accurate sub-models across various architectures and tasks."