File Download

There are no files associated with this item.

  Links for fulltext
     (May Require Subscription)
Supplementary

Article: FedBM: Stealing knowledge from pre-trained language models for heterogeneous federated learning

TitleFedBM: Stealing knowledge from pre-trained language models for heterogeneous federated learning
Authors
KeywordsFederated learning
Medical Image Classification
Pre-trained language model
Issue Date1-May-2025
PublisherElsevier
Citation
Medical Image Analysis, 2025, v. 102 How to Cite?
Abstract

Federated learning (FL) has shown great potential in medical image computing since it provides a decentralized learning paradigm that allows multiple clients to train a model collaboratively without privacy leakage. However, current studies have shown that data heterogeneity incurs local learning bias in classifiers and feature extractors of client models during local training, leading to the performance degradation of a federation system. To address these issues, we propose a novel framework called Federated Bias eliMinating (FedBM) to get rid of local learning bias in heterogeneous federated learning (FL), which mainly consists of two modules, i.e., Linguistic Knowledge-based Classifier Construction (LKCC) and Concept-guided Global Distribution Estimation (CGDE). Specifically, LKCC exploits class concepts, prompts and pre-trained language models (PLMs) to obtain concept embeddings. These embeddings are used to estimate the latent concept distribution of each class in the linguistic space. Based on the theoretical derivation, we can rely on these distributions to pre-construct a high-quality classifier for clients to achieve classification optimization, which is frozen to avoid classifier bias during local training. CGDE samples probabilistic concept embeddings from the latent concept distributions to learn a conditional generator to capture the input space of the global model. Three regularization terms are introduced to improve the quality and utility of the generator. The generator is shared by all clients and produces pseudo data to calibrate updates of local feature extractors. Extensive comparison experiments and ablation studies on public datasets demonstrate the superior performance of FedBM over state-of-the-arts and confirm the effectiveness of each module, respectively. The code is available at https://github.com/CUHK-AIM-Group/FedBM.


Persistent Identifierhttp://hdl.handle.net/10722/360482
ISSN
2023 Impact Factor: 10.7
2023 SCImago Journal Rankings: 4.112

 

DC FieldValueLanguage
dc.contributor.authorZhu, Meilu-
dc.contributor.authorYang, Qiushi-
dc.contributor.authorGao, Zhifan-
dc.contributor.authorYuan, Yixuan-
dc.contributor.authorLiu, Jun-
dc.date.accessioned2025-09-11T00:30:41Z-
dc.date.available2025-09-11T00:30:41Z-
dc.date.issued2025-05-01-
dc.identifier.citationMedical Image Analysis, 2025, v. 102-
dc.identifier.issn1361-8415-
dc.identifier.urihttp://hdl.handle.net/10722/360482-
dc.description.abstract<p>Federated learning (FL) has shown great potential in medical image computing since it provides a decentralized learning paradigm that allows multiple clients to train a model collaboratively without privacy leakage. However, current studies have shown that data heterogeneity incurs local learning bias in classifiers and feature extractors of client models during local training, leading to the performance degradation of a federation system. To address these issues, we propose a novel framework called Federated Bias eliMinating (FedBM) to get rid of local learning bias in heterogeneous federated learning (FL), which mainly consists of two modules, i.e., Linguistic Knowledge-based Classifier Construction (LKCC) and Concept-guided Global Distribution Estimation (CGDE). Specifically, LKCC exploits class concepts, prompts and pre-trained language models (PLMs) to obtain concept embeddings. These embeddings are used to estimate the latent concept distribution of each class in the linguistic space. Based on the theoretical derivation, we can rely on these distributions to pre-construct a high-quality classifier for clients to achieve classification optimization, which is frozen to avoid classifier bias during local training. CGDE samples probabilistic concept embeddings from the latent concept distributions to learn a conditional generator to capture the input space of the global model. Three regularization terms are introduced to improve the quality and utility of the generator. The generator is shared by all clients and produces pseudo data to calibrate updates of local feature extractors. Extensive comparison experiments and ablation studies on public datasets demonstrate the superior performance of FedBM over state-of-the-arts and confirm the effectiveness of each module, respectively. The code is available at https://github.com/CUHK-AIM-Group/FedBM.</p>-
dc.languageeng-
dc.publisherElsevier-
dc.relation.ispartofMedical Image Analysis-
dc.subjectFederated learning-
dc.subjectMedical Image Classification-
dc.subjectPre-trained language model-
dc.titleFedBM: Stealing knowledge from pre-trained language models for heterogeneous federated learning-
dc.typeArticle-
dc.identifier.doi10.1016/j.media.2025.103524-
dc.identifier.pmid40073584-
dc.identifier.scopuseid_2-s2.0-86000465365-
dc.identifier.volume102-
dc.identifier.issnl1361-8415-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats