Skip to main content

Rethinking normalization methods in federated learning

Publication ,  Conference
Du, Z; Sun, J; Li, A; Chen, PY; Zhang, J; Li, H; Chen, Y
Published in: DistributedML 2022 - Proceedings of the 3rd International Workshop on Distributed Machine Learning, Part of CoNEXT 2022
December 9, 2022

Federated learning (FL) is a popular distributed learning framework that can reduce privacy risks by not explicitly sharing private data. In this work, we explicitly uncover external covariate shift problem in FL, which is caused by the independent local training processes on different devices. We demonstrate that external covariate shifts will lead to the obliteration of some devices' contributions to the global model. Further, we show that normalization layers are indispensable in FL since their inherited properties can alleviate the problem of obliterating some devices' contributions. However, recent works have shown that batch normalization, which is one of the standard components in many deep neural networks, will incur accuracy drop of the global model in FL. The essential reason for the failure of batch normalization in FL is poorly studied. We unveil that external covariate shift is the key reason why batch normalization is ineffective in FL. We also show that layer normalization is a better choice in FL which can mitigate the external covariate shift and improve the performance of the global model. We conduct experiments on CIFAR10 under non-IID settings. The results demonstrate that models with layer normalization converge fastest and achieve the best or comparable accuracy for three different model architectures.

Duke Scholars

Altmetric Attention Stats
Dimensions Citation Stats

Published In

DistributedML 2022 - Proceedings of the 3rd International Workshop on Distributed Machine Learning, Part of CoNEXT 2022

DOI

ISBN

9781450399227

Publication Date

December 9, 2022

Start / End Page

16 / 22
 

Citation

APA
Chicago
ICMJE
MLA
NLM
Du, Z., Sun, J., Li, A., Chen, P. Y., Zhang, J., Li, H., & Chen, Y. (2022). Rethinking normalization methods in federated learning. In DistributedML 2022 - Proceedings of the 3rd International Workshop on Distributed Machine Learning, Part of CoNEXT 2022 (pp. 16–22). https://doi.org/10.1145/3565010.3569062
Du, Z., J. Sun, A. Li, P. Y. Chen, J. Zhang, H. Li, and Y. Chen. “Rethinking normalization methods in federated learning.” In DistributedML 2022 - Proceedings of the 3rd International Workshop on Distributed Machine Learning, Part of CoNEXT 2022, 16–22, 2022. https://doi.org/10.1145/3565010.3569062.
Du Z, Sun J, Li A, Chen PY, Zhang J, Li H, et al. Rethinking normalization methods in federated learning. In: DistributedML 2022 - Proceedings of the 3rd International Workshop on Distributed Machine Learning, Part of CoNEXT 2022. 2022. p. 16–22.
Du, Z., et al. “Rethinking normalization methods in federated learning.” DistributedML 2022 - Proceedings of the 3rd International Workshop on Distributed Machine Learning, Part of CoNEXT 2022, 2022, pp. 16–22. Scopus, doi:10.1145/3565010.3569062.
Du Z, Sun J, Li A, Chen PY, Zhang J, Li H, Chen Y. Rethinking normalization methods in federated learning. DistributedML 2022 - Proceedings of the 3rd International Workshop on Distributed Machine Learning, Part of CoNEXT 2022. 2022. p. 16–22.

Published In

DistributedML 2022 - Proceedings of the 3rd International Workshop on Distributed Machine Learning, Part of CoNEXT 2022

DOI

ISBN

9781450399227

Publication Date

December 9, 2022

Start / End Page

16 / 22