FedDAR: Federated Learning With Data-Quantity Aware Regularization for Heterogeneous Distributed Data
Federated learning (FL) has emerged as a promising approach for collaboratively training global models and classifiers without sharing private data. However, existing studies primarily focus on distinct methodologies for typical and personalized FL (tFL and pFL), representing a challenge in explorin...
Saved in:
| Main Authors: | , |
|---|---|
| Format: | Article |
| Language: | English |
| Published: |
IEEE
2025-01-01
|
| Series: | IEEE Access |
| Subjects: | |
| Online Access: | https://ieeexplore.ieee.org/document/11091276/ |
| Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
| _version_ | 1849240550077628416 |
|---|---|
| author | Youngjun Kwak Minyoung Jung |
| author_facet | Youngjun Kwak Minyoung Jung |
| author_sort | Youngjun Kwak |
| collection | DOAJ |
| description | Federated learning (FL) has emerged as a promising approach for collaboratively training global models and classifiers without sharing private data. However, existing studies primarily focus on distinct methodologies for typical and personalized FL (tFL and pFL), representing a challenge in exploring cross-applicable training methods. Moreover, previous approaches often rely on data and feature augmentation branches, overlooking data-quantity considerations, leading to suboptimal performance and inefficient communication costs, particularly in multi-class classification tasks. To address these challenges, we propose a novel add-on regularization technique for existing FL methods, named Data-quantity Aware Regularization (FedDAR), seamlessly integrating with existing tFL and pFL frameworks. This network-agnostic methodology reformulates the local training procedure by incorporating two crucial components: 1) enriched-feature augmentation, where features of the local model are coordinated with pre-initialized features to ensure unbiased-representations with efficient global communication rounds for unbalanced data distribution, and 2) data-quantity aware branch, which associates with local data size to improve the optimization of the local model using both supervised and self-supervised labels. We demonstrate significant performance improvements in tFL and pFL, achieving state-of-the-art results across MNIST, F-MNIST, CIFAR-10/100, and Tiny-ImageNet benchmarks. |
| format | Article |
| id | doaj-art-d5ed5f3b19d4472b9df8a4c36cc31c36 |
| institution | Kabale University |
| issn | 2169-3536 |
| language | English |
| publishDate | 2025-01-01 |
| publisher | IEEE |
| record_format | Article |
| series | IEEE Access |
| spelling | doaj-art-d5ed5f3b19d4472b9df8a4c36cc31c362025-08-20T04:00:33ZengIEEEIEEE Access2169-35362025-01-011313320813321710.1109/ACCESS.2025.359183911091276FedDAR: Federated Learning With Data-Quantity Aware Regularization for Heterogeneous Distributed DataYoungjun Kwak0https://orcid.org/0009-0004-4805-8786Minyoung Jung1https://orcid.org/0009-0001-6995-5687Financial Tech Laboratory, KakaoBank Corporation, Seongnam-si, South KoreaKETI, Korea Electronics Technology Institute, Seongnam-si, South KoreaFederated learning (FL) has emerged as a promising approach for collaboratively training global models and classifiers without sharing private data. However, existing studies primarily focus on distinct methodologies for typical and personalized FL (tFL and pFL), representing a challenge in exploring cross-applicable training methods. Moreover, previous approaches often rely on data and feature augmentation branches, overlooking data-quantity considerations, leading to suboptimal performance and inefficient communication costs, particularly in multi-class classification tasks. To address these challenges, we propose a novel add-on regularization technique for existing FL methods, named Data-quantity Aware Regularization (FedDAR), seamlessly integrating with existing tFL and pFL frameworks. This network-agnostic methodology reformulates the local training procedure by incorporating two crucial components: 1) enriched-feature augmentation, where features of the local model are coordinated with pre-initialized features to ensure unbiased-representations with efficient global communication rounds for unbalanced data distribution, and 2) data-quantity aware branch, which associates with local data size to improve the optimization of the local model using both supervised and self-supervised labels. We demonstrate significant performance improvements in tFL and pFL, achieving state-of-the-art results across MNIST, F-MNIST, CIFAR-10/100, and Tiny-ImageNet benchmarks.https://ieeexplore.ieee.org/document/11091276/Federated learningregularizationclassificationaugmentationcommunication efficiencydata-quantity awareness |
| spellingShingle | Youngjun Kwak Minyoung Jung FedDAR: Federated Learning With Data-Quantity Aware Regularization for Heterogeneous Distributed Data IEEE Access Federated learning regularization classification augmentation communication efficiency data-quantity awareness |
| title | FedDAR: Federated Learning With Data-Quantity Aware Regularization for Heterogeneous Distributed Data |
| title_full | FedDAR: Federated Learning With Data-Quantity Aware Regularization for Heterogeneous Distributed Data |
| title_fullStr | FedDAR: Federated Learning With Data-Quantity Aware Regularization for Heterogeneous Distributed Data |
| title_full_unstemmed | FedDAR: Federated Learning With Data-Quantity Aware Regularization for Heterogeneous Distributed Data |
| title_short | FedDAR: Federated Learning With Data-Quantity Aware Regularization for Heterogeneous Distributed Data |
| title_sort | feddar federated learning with data quantity aware regularization for heterogeneous distributed data |
| topic | Federated learning regularization classification augmentation communication efficiency data-quantity awareness |
| url | https://ieeexplore.ieee.org/document/11091276/ |
| work_keys_str_mv | AT youngjunkwak feddarfederatedlearningwithdataquantityawareregularizationforheterogeneousdistributeddata AT minyoungjung feddarfederatedlearningwithdataquantityawareregularizationforheterogeneousdistributeddata |