File Download
There are no files associated with this item.
Links for fulltext
(May Require Subscription)
- Publisher Website: 10.1109/CVPR52688.2022.00982
- Scopus: eid_2-s2.0-85138843349
- WOS: WOS:000870759103013
- Find via
Supplementary
- Citations:
- Appears in Collections:
Conference Paper: Rethinking Architecture Design for Tackling Data Heterogeneity in Federated Learning
Title | Rethinking Architecture Design for Tackling Data Heterogeneity in Federated Learning |
---|---|
Authors | |
Keywords | Privacy and federated learning |
Issue Date | 2022 |
Citation | Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 2022, v. 2022-June, p. 10051-10061 How to Cite? |
Abstract | Federated learning is an emerging research paradigm enabling collaborative training of machine learning models among different organizations while keeping data private at each institution. Despite recent progress, there remain fundamental challenges such as the lack of convergence and the potential for catastrophic forgetting across real-world heterogeneous devices. In this paper, we demonstrate that self-attention-based architectures (e.g., Transformers) are more robust to distribution shifts and hence improve federated learning over heterogeneous data. Concretely, we conduct the first rigorous empirical investigation of different neural architectures across a range of federated algorithms, real-world benchmarks, and heterogeneous data splits. Our experiments show that simply replacing convolutional networks with Transformers can greatly reduce catastrophic forgetting of previous devices, accelerate convergence, and reach a better global model, especially when dealing with heterogeneous data. We release our code and pretrained models to encourage future exploration in robust architectures as an alternative to current research efforts on the optimization front. |
Persistent Identifier | http://hdl.handle.net/10722/325581 |
ISSN | 2023 SCImago Journal Rankings: 10.331 |
ISI Accession Number ID |
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Qu, Liangqiong | - |
dc.contributor.author | Zhou, Yuyin | - |
dc.contributor.author | Liang, Paul Pu | - |
dc.contributor.author | Xia, Yingda | - |
dc.contributor.author | Wang, Feifei | - |
dc.contributor.author | Adeli, Ehsan | - |
dc.contributor.author | Fei-Fei, Li | - |
dc.contributor.author | Rubin, Daniel | - |
dc.date.accessioned | 2023-02-27T07:34:32Z | - |
dc.date.available | 2023-02-27T07:34:32Z | - |
dc.date.issued | 2022 | - |
dc.identifier.citation | Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 2022, v. 2022-June, p. 10051-10061 | - |
dc.identifier.issn | 1063-6919 | - |
dc.identifier.uri | http://hdl.handle.net/10722/325581 | - |
dc.description.abstract | Federated learning is an emerging research paradigm enabling collaborative training of machine learning models among different organizations while keeping data private at each institution. Despite recent progress, there remain fundamental challenges such as the lack of convergence and the potential for catastrophic forgetting across real-world heterogeneous devices. In this paper, we demonstrate that self-attention-based architectures (e.g., Transformers) are more robust to distribution shifts and hence improve federated learning over heterogeneous data. Concretely, we conduct the first rigorous empirical investigation of different neural architectures across a range of federated algorithms, real-world benchmarks, and heterogeneous data splits. Our experiments show that simply replacing convolutional networks with Transformers can greatly reduce catastrophic forgetting of previous devices, accelerate convergence, and reach a better global model, especially when dealing with heterogeneous data. We release our code and pretrained models to encourage future exploration in robust architectures as an alternative to current research efforts on the optimization front. | - |
dc.language | eng | - |
dc.relation.ispartof | Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition | - |
dc.subject | Privacy and federated learning | - |
dc.title | Rethinking Architecture Design for Tackling Data Heterogeneity in Federated Learning | - |
dc.type | Conference_Paper | - |
dc.description.nature | link_to_subscribed_fulltext | - |
dc.identifier.doi | 10.1109/CVPR52688.2022.00982 | - |
dc.identifier.scopus | eid_2-s2.0-85138843349 | - |
dc.identifier.volume | 2022-June | - |
dc.identifier.spage | 10051 | - |
dc.identifier.epage | 10061 | - |
dc.identifier.isi | WOS:000870759103013 | - |