Alapfogalmak
ProgFed is a novel federated learning framework that reduces communication and computation costs by progressively training increasingly complex models, achieving comparable or superior performance to traditional methods.
Statisztikák
ProgFed saves up to 20% computation and up to 63% communication costs for converged models.
The approach can achieve a wide range of trade-offs by combining these techniques, showing reduced communication of up to 50x at only 0.1% loss in utility.
ProgFed saves around 25% computation cost and up to 32% two-way communication costs in federated classification.
In federated segmentation, ProgFed reduces communication costs by 63% without sacrificing performance.
ProgFed allows for a communication cost reduction of around 2x in classification and 6.5x in U-net segmentation while achieving practicable performance (≥98% of the best baseline).
Idézetek
"We propose ProgFed, the first federated progressive learning framework that reduces both communication and computation costs while preserving model utility."
"Our method inherently reduces two-way communication costs and complements existing methods."
"ProgFed is compatible with classical compression, including sparsification and quantization, and various federated optimizations, such as FedAvg, FedProx, and FedAdam."