Fed-Hetero: A Self-Evaluating Federated Learning Framework for Data Heterogeneity

Federated learning (FL) enables deep learning models to be trained locally on devices without the need for data sharing, ensuring data privacy. However, when clients have uneven or imbalanced data distributions, it leads to data heterogeneity. Data heterogeneity can appear in different ways, often d...

Full description

Saved in:
Bibliographic Details
Main Authors: Aiswariya Milan Kummaya, Amudha Joseph, Kumar Rajamani, George Ghinea
Format: Article
Language:English
Published: MDPI AG 2025-02-01
Series:Applied System Innovation
Subjects:
Online Access:https://www.mdpi.com/2571-5577/8/2/28
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:Federated learning (FL) enables deep learning models to be trained locally on devices without the need for data sharing, ensuring data privacy. However, when clients have uneven or imbalanced data distributions, it leads to data heterogeneity. Data heterogeneity can appear in different ways, often due to variations in label, data distributions, feature variations, and structural inconsistencies in the images. This can significantly impact FL performance, as the global model often struggles to achieve optimal convergence. To enhance training efficiency and model performance, a common strategy in FL is to exclude clients with limited data. However, excluding such clients can raise fairness concerns, particularly for smaller populations. To understand the influence of data heterogeneity, a self-evaluating federated learning framework for heterogeneity, <b>Fed-Hetero,</b> was designed to assess the type of heterogeneity associated with the clients and provide recommendations to clients to enhance the global model’s accuracy. Fed-Hetero thus enables the clients with limited data to participate in FL processes by adopting appropriate strategies that enhance model accuracy. The results show that Fed-Hetero identifies the client with heterogeneity and provides personalized recommendations.
ISSN:2571-5577