Client Selection in Federated Learning under Imperfections in Environment
Permanent lenke
https://hdl.handle.net/10037/28341Dato
2022-02-25Type
Journal articleTidsskriftartikkel
Peer reviewed
Sammendrag
Federated learning promises an elegant solution for learning global models across distributed and privacy-protected datasets. However, challenges related to skewed data distribution, limited computational and communication resources, data poisoning, and free riding clients affect the performance of federated learning. Selection of the best clients for each round of learning is critical in alleviating these problems. We propose a novel sampling method named the irrelevance sampling technique. Our method is founded on defining a novel irrelevance score that incorporates the client characteristics in a single floating value, which can elegantly classify the client into three numerical sign defined pools for easy sampling. It is a computationally inexpensive, intuitive and privacy preserving sampling technique that selects a subset of clients based on quality and quantity of data on edge devices. It achieves 50–80% faster convergence even in highly skewed data distribution in the presence of free riders based on lack of data and severe class imbalance under both Independent and Identically Distributed (IID) and Non-IID conditions. It shows good performance on practical application datasets.
Forlag
MDPISitering
Kumari, Rai, Prasad DK. Client Selection in Federated Learning under Imperfections in Environment. AI. 2022;3(1):124-145Metadata
Vis full innførselSamlinger
Copyright 2022 The Author(s)