מושגי ליבה
Heterogeneous data in federated learning leads to dimensional collapse in both global and local models, which can be effectively mitigated by the proposed FEDDECORR method.
תקציר
The paper studies the impact of data heterogeneity on the representations learned in federated learning. Key observations and insights:
- Empirical observations:
- Stronger data heterogeneity among clients leads to more severe dimensional collapse in the representations of both global and local models.
- The dimensional collapse of the global model is inherited from the local models.
- Theoretical analysis:
- Analyzes the gradient flow dynamics of local training and shows that heterogeneous data drives the weight matrices of local models to be biased towards low-rank, resulting in dimensional collapse of representations.
- Proposed method - FEDDECORR:
- Adds a regularization term during local training to encourage the representations to be decorrelated, effectively mitigating dimensional collapse.
- FEDDECORR is computationally efficient and can be easily integrated with existing federated learning methods.
- Experiments:
- FEDDECORR consistently outperforms baseline federated learning methods across various datasets and heterogeneity settings.
- The improvements from FEDDECORR become more pronounced as the number of clients increases, demonstrating its effectiveness in large-scale federated learning.
סטטיסטיקה
"As the degree of data heterogeneity increases, more singular values tend to evolve towards zero."
"Heterogeneous data drive the weight matrices of the local models to be biased to being low-rank, which further results in representation dimensional collapse."
ציטוטים
"Interestingly, we find that as the degree of data heterogeneity increases, more singular values tend to evolve towards zero. This observation suggests that stronger data heterogeneity causes the trained global model to suffer from more severe dimensional collapse, whereby representations are biased towards residing in a lower-dimensional space (or manifold)."
"Essentially, dimensional collapse is a form of oversimplification in terms of the model, where the representation space is not being fully utilized to discriminate diverse data of different classes."