Core Concepts
The pMAE method leverages the reconstruction capabilities of masked autoencoders (MAEs) and prompt tuning to address catastrophic forgetting and non-IID issues in federated continual learning, achieving parameter efficiency and improved performance compared to existing prompt-based methods.
Stats
The CUB-200 dataset comprises approximately 12,000 images spanning 200 fine-grained bird categories.
The ImageNet-R dataset includes 200 classes with 24,000 training images and 6,000 testing images.
Experiments were conducted with T = 20 tasks, each containing 10 classes.
The number of clients was set to K = 10.
A total of Rall = 200 communication rounds and local update epochs E = 5 were used.
For pMAE, the uploaded number of restore information was set to u = 4, and fine-tuning on the server was performed over Eserver = 5 epochs.
The masking ratio for MAE is 75%.
The discriminative prompt length is set to Lp = 20 and is inserted into the first five transformer blocks.
The reconstructive prompt is set with Lp = 5 and is only inserted into the first layer.