The paper proposes FedStitch, a novel federated learning (FL) framework that addresses the memory and energy limitations of deploying FL on resource-constrained devices. Unlike traditional approaches that train the global model from scratch, FedStitch composes the global model by stitching together pre-trained blocks from diverse neural network models.
The key highlights of FedStitch are:
Initialization: The server divides pre-trained models into blocks and distributes them to participating clients as a candidate pool.
Local Block Selection: Each client selects the most suitable block from the pool based on the compatibility with their local data, measured by Centered Kernel Alignment (CKA) scores.
Weighted Aggregation: The server uses a reinforcement learning-based weighted aggregation to select the optimal blocks, mitigating the impact of non-IID data distribution across clients.
Search Space Optimization: The server continuously reduces the size of the candidate block pool during the stitching process to accelerate the overall generation.
Local Energy Optimization: Each client employs a feedback-based frequency configuration method to minimize energy consumption while meeting the server's deadlines.
The experiments demonstrate that FedStitch significantly improves model accuracy by up to 20.93% compared to existing approaches. It also achieves up to 8.12x speedup, reduces memory footprint by up to 79.5%, and saves up to 89.41% energy during the learning procedure.
To Another Language
from source content
arxiv.org
Key Insights Distilled From
by Shichen Zhan... at arxiv.org 09-12-2024
https://arxiv.org/pdf/2409.07202.pdfDeeper Inquiries