The paper addresses the communication and energy consumption bottlenecks in federated learning (FL) by proposing a zero-order (ZO) optimization method. In the standard FL approach, each device computes the gradient of the local loss function and sends it to the server, which leads to high communication overhead, especially when the model has a large number of parameters.
The proposed method, called digital zero-order federated learning (DZOFL), avoids the computation and exchange of gradients. Instead, each device queries its local loss function twice with a random perturbation and sends the quantized difference of the two queries to the server. The server aggregates the received scalars and sends the quantized aggregated scalar back to the devices, which then update the model.
The key advantages of the DZOFL method are:
The paper provides a detailed convergence analysis of the proposed method in the non-convex setting, considering the impact of quantization and packet dropping due to wireless errors. It is shown that the method achieves a convergence rate that competes with standard gradient-based FL techniques while requiring much less communication overhead.
Numerical results demonstrate that the DZOFL method outperforms the standard FL approach in terms of convergence time and energy consumption.
toiselle kielelle
lähdeaineistosta
arxiv.org
Tärkeimmät oivallukset
by Elissa Mhann... klo arxiv.org 09-26-2024
https://arxiv.org/pdf/2409.16456.pdfSyvällisempiä Kysymyksiä