핵심 개념
Enhanced transformer architecture ParFormer integrates token mixers for improved feature extraction capabilities.
통계
Our model variants with 11M, 23M, and 34M parameters achieve scores of 80.4%, 82.1%, and 83.1%, respectively.
인용구
"Our ParFormer outperforms CNN-based and state-of-the-art transformer-based architectures in image classification."
"The proposed CAPE has been demonstrated to benefit the overall MetaFormer architecture."