Core Concepts
Hierarchical Federated Learning with Hierarchical Differential Privacy (H2FDP) is a methodology that jointly optimizes privacy and performance in hierarchical networks by adapting differential privacy noise injection at different layers of the established federated learning hierarchy.
Abstract
The paper proposes Hierarchical Federated Learning with Hierarchical Differential Privacy (H2FDP), a framework that integrates flexible Hierarchical Differential Privacy (HDP) trust models with hierarchical federated learning (HFL) to preserve privacy throughout the entire training process.
Key highlights:
H2FDP adapts the differential privacy (DP) noise injection at different layers of the HFL hierarchy (edge devices, edge servers, cloud server) based on the trust models within particular subnetworks.
The authors provide a comprehensive convergence analysis of H2FDP, revealing conditions on parameter tuning under which the training process converges sublinearly to a finite stationarity gap that depends on the network hierarchy, trust model, and target privacy level.
Leveraging the convergence analysis, the authors develop an adaptive control algorithm for H2FDP that tunes local model training properties to minimize communication energy, latency, and the stationarity gap while maintaining a sub-linear convergence rate and desired privacy criteria.
Numerical evaluations demonstrate that H2FDP obtains substantial improvements in these metrics over baselines for different privacy budgets, and validate the impact of different system configurations.
Stats
The 𝐿2-norm sensitivity of the exchanged gradients during local aggregations is 2𝜂𝑘𝜏𝑘𝐺/𝑠𝑐 for secure edge servers and 2𝜂𝑘𝜏𝑘𝐺 for insecure edge servers.
The 𝐿2-norm sensitivity of the exchanged gradients during global aggregations is 2𝜂𝑘𝜏𝑘𝐺/𝑠𝑐 for secure edge servers and 2𝜂𝑘𝜏𝑘𝐺 for insecure edge servers.
Quotes
"Hierarchical Federated Learning with Hierarchical Differential Privacy (H2FDP) is a DP-enhanced FL methodology for jointly optimizing privacy and performance in hierarchical networks."
"Our convergence analysis (culminating in Theorem 4.3) shows that with an appropriate choice of FL step size, the cumulative average global model will converge sublinearly with rate O(1/√k) to a region around a stationary point."