Chrome Extension
WeChat Mini Program
Use on ChatGLM

Towards Scalable Resilient Federated Learning: A Fully Decentralised Approach.

PerCom Workshops(2023)

Cited 0|Views6
No score
Abstract
Federated Learning (FL) collaboratively trains machine learning models on the data of local devices without having to move the data itself: a central server aggregates models, with privacy and performance benefits but also scalability and resilience challenges. In this paper we present FDFL, a new fully decentralized FL model and architecture that improves standard FL scalability and resilience with no loss of convergence speed. FDFL provides an aggregator-based model that enables scalability benefits and features an election process to tolerate node failures. Simulation results show that FDFL scales well with network size in terms of computing, memory, and communication compared to related FL approaches such as standard FL, FL with aggregators, or FL with election, with also good resilience to node failures.
More
Translated text
Key words
Federated learning,decentralized learning,pervasive machine learning,edge AI,scalability,resilience
AI Read Science
Must-Reading Tree
Example
Generate MRT to find the research sequence of this paper
Chat Paper
Summary is being generated by the instructions you defined