Yuan, J., Wei, Z., & Guo, W. (2024). Mixed-Precision Federated Learning via Multi-Precision Over-The-Air Aggregation. arXiv preprint arXiv:2406.03402v2.
This paper investigates the potential of using heterogeneous quantization levels in federated learning (FL) to address computational and communication limitations, focusing on the benefits of mixed-precision clients compared to homogeneous precision setups.
The authors propose a novel mixed-precision OTA-FL framework that utilizes a multi-precision gradient modulation scheme for over-the-air aggregation, eliminating the need for precision conversion. They evaluate their approach through simulations using the German Traffic Sign Recognition Benchmark (GTSRB) dataset, comparing various client quantization schemes and analyzing server and client performance metrics, including convergence speed, accuracy, and energy consumption.
The study demonstrates that mixed-precision OTA-FL, employing heterogeneous client quantization levels and a tailored aggregation scheme, effectively balances performance and energy efficiency in federated learning. The framework proves particularly advantageous in resource-constrained edge computing environments by enabling the participation of ultra-low precision clients without compromising overall system accuracy.
This research provides valuable insights into optimizing federated learning for heterogeneous hardware environments, paving the way for more efficient and scalable deployments of FL in real-world applications with diverse resource constraints.
The current study focuses on a simulated environment. Future research should explore the practical implementation and evaluation of the proposed framework in real-world settings with varying network conditions and device capabilities. Additionally, investigating more sophisticated quantization schemes and exploring the trade-offs between accuracy, energy efficiency, and communication overhead could further enhance the framework's effectiveness.
Іншою мовою
із вихідного контенту
arxiv.org
Ключові висновки, отримані з
by Jinsheng Yua... о arxiv.org 10-17-2024
https://arxiv.org/pdf/2406.03402.pdfГлибші Запити