We are in an early-release beta. Expect some adventures and rough edges.
Federated Learning (FL) has emerged as a practical paradigm for collaborative training under privacy and communication constraints. However, FL faces various challenges due to the diversity of edge devices. In addition to differences in computing performance and communication efficiency, data distributed on these devices is often imbalanced. While there have been numerous studies on these problems, many introduce additional hyperparameters and communication costs. To address these challenges, we propose an efficient FL method based on virtual clients, a new training framework we have developed. Our approach fills the gap between balanced and imbalanced data distribution by using virtual clients. Our algorithm outperforms classical FedAvg and the state-of-the-art for non-IID data in extensive experiments. Particularly in highly heterogeneous settings, FedVC demonstrates significantly more stable and accurate convergence behavior than other approaches.
The approach to dividing data in federated learning is diverse. Assuming that there is an optimal data division,
To begin, let's consider the
As shown in the following figures, FedVC demonstrates more stable and accurate convergence behavior than other methods.
This repository's code is based on the PaddlePaddle framework. To run it, you should first import it into the aistudio environment.