Core Concepts
GPT-NAS leverages the pattern recognition and generative capabilities of pre-trained GPT models to enhance the efficiency of evolutionary algorithms in finding optimal neural architectures.
Stats
GPT-NAS achieves 97.69% accuracy on CIFAR-10, outperforming the manually designed ResNet-101 by 4.12%.
On CIFAR-100, GPT-NAS achieves 82.81% accuracy, surpassing other EA-NAS algorithms by a significant margin.
GPT-NAS achieves 79.08% Top-1 accuracy and 95.92% Top-5 accuracy on ImageNet-1K, surpassing all other compared algorithms.
Introducing the GPT model in the NAS process led to accuracy improvements of 7%, 9%, and 12% on CIFAR-10, CIFAR-100, and ImageNet-1K, respectively.
Quotes
"While neural architectures have achieved human-level performances in several tasks, only a few of them have been obtained from the NAS method."
"The main challenge with NAS is that its effectiveness is often hindered by the vast search space of possible architectures."
"To this end, we propose a NAS algorithm based on Generative Pre-trained Transformer (GPT-NAS), which is an innovative solution for the large search space."
"Unlike traditional approaches that focus solely on the search space or search strategy, our proposed GPT-NAS algorithm leverages the power of GPT [23] models to introduce a priori knowledge into the algorithm."