DRAG: Divergence-based Adaptive Aggregation in Federated learning on Non-IID Data

09/04/2023
by   Feng Zhu, et al.
1

Local stochastic gradient descent (SGD) is a fundamental approach in achieving communication efficiency in Federated Learning (FL) by allowing individual workers to perform local updates. However, the presence of heterogeneous data distributions across working nodes causes each worker to update its local model towards a local optimum, leading to the phenomenon known as “client-drift" and resulting in slowed convergence. To address this issue, previous works have explored methods that either introduce communication overhead or suffer from unsteady performance. In this work, we introduce a novel metric called “degree of divergence," quantifying the angle between the local gradient and the global reference direction. Leveraging this metric, we propose the divergence-based adaptive aggregation (DRAG) algorithm, which dynamically “drags" the received local updates toward the reference direction in each round without requiring extra communication overhead. Furthermore, we establish a rigorous convergence analysis for DRAG, proving its ability to achieve a sublinear convergence rate. Compelling experimental results are presented to illustrate DRAG's superior performance compared to state-of-the-art algorithms in effectively managing the client-drift phenomenon. Additionally, DRAG exhibits remarkable resilience against certain Byzantine attacks. By securely sharing a small sample of the client's data with the FL server, DRAG effectively counters these attacks, as demonstrated through comprehensive experiments.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/19/2022

Adaptive Control of Client Selection and Gradient Compression for Efficient Federated Learning

Federated learning (FL) allows multiple clients cooperatively train mode...
research
07/19/2023

FedBug: A Bottom-Up Gradual Unfreezing Framework for Federated Learning

Federated Learning (FL) offers a collaborative training framework, allow...
research
10/06/2022

Communication-Efficient and Drift-Robust Federated Learning via Elastic Net

Federated learning (FL) is a distributed method to train a global model ...
research
10/28/2022

Aggregation in the Mirror Space (AIMS): Fast, Accurate Distributed Machine Learning in Military Settings

Distributed machine learning (DML) can be an important capability for mo...
research
08/20/2021

Accelerating Federated Learning with a Global Biased Optimiser

Federated Learning (FL) is a recent development in the field of machine ...
research
06/09/2023

Understanding How Consistency Works in Federated Learning via Stage-wise Relaxed Initialization

Federated learning (FL) is a distributed paradigm that coordinates massi...
research
02/12/2020

Federated Clustering via Matrix Factorization Models: From Model Averaging to Gradient Sharing

Recently, federated learning (FL) has drawn significant attention due to...

Please sign up or login with your details

Forgot password? Click here to reset