DeepAI AI Chat
Log In Sign Up

Communication-Efficient Online Federated Learning Framework for Nonlinear Regression

by   Vinay Chakravarthi Gogineni, et al.
University of Notre Dame
University of Hawaii

Federated learning (FL) literature typically assumes that each client has a fixed amount of data, which is unrealistic in many practical applications. Some recent works introduced a framework for online FL (Online-Fed) wherein clients perform model learning on streaming data and communicate the model to the server; however, they do not address the associated communication overhead. As a solution, this paper presents a partial-sharing-based online federated learning framework (PSO-Fed) that enables clients to update their local models using continuous streaming data and share only portions of those updated models with the server. During a global iteration of PSO-Fed, non-participant clients have the privilege to update their local models with new data. Here, we consider a global task of kernel regression, where clients use a random Fourier features-based kernel LMS on their data for local learning. We examine the mean convergence of the PSO-Fed for kernel regression. Experimental results show that PSO-Fed can achieve competitive performance with a significantly lower communication overhead than Online-Fed.


page 1

page 2

page 3

page 4


Resource-Aware Asynchronous Online Federated Learning for Nonlinear Regression

Many assumptions in the federated learning literature present a best-cas...

Multiple Kernel-Based Online Federated Learning

Online federated learning (OFL) becomes an emerging learning framework, ...

Asynchronous Online Federated Learning with Reduced Communication Requirements

Online federated learning (FL) enables geographically distributed device...

GRP-FED: Addressing Client Imbalance in Federated Learning via Global-Regularized Personalization

Since data is presented long-tailed in reality, it is challenging for Fe...

FedHB: Hierarchical Bayesian Federated Learning

We propose a novel hierarchical Bayesian approach to Federated Learning ...

OFedQIT: Communication-Efficient Online Federated Learning via Quantization and Intermittent Transmission

Online federated learning (OFL) is a promising framework to collaborativ...

Federated Learning under Heterogeneous and Correlated Client Availability

The enormous amount of data produced by mobile and IoT devices has motiv...