Open Access iconOpen Access

ARTICLE

crossmark

A Client Selection Method Based on Loss Function Optimization for Federated Learning

Yan Zeng1,2,3, Siyuan Teng1, Tian Xiang4,*, Jilin Zhang1,2,3, Yuankai Mu5, Yongjian Ren1,2,3,*, Jian Wan1,2,3

1 School of Computer Science and Technology, Hangzhou Dianzi University, Hangzhou, 310018, China
2 Key Laboratory of Complex System Modeling and Simulation, Ministry of Education, Hangzhou, 310018, China
3 Zhejiang Engineering Research Center of Data Security Governance, Hangzhou, 310018, China
4 Intelligent Robotics Research Center, Zhejiang Lab, Hangzhou, 311100, China
5 HDU-ITMO Joint Institute, Hangzhou Dianzi University, Hangzhou, 310018, China

* Corresponding Authors: Tian Xiang. Email: email; Yongjian Ren. Email: email

(This article belongs to the Special Issue: Federated Learning Algorithms, Approaches, and Systems for Internet of Things)

Computer Modeling in Engineering & Sciences 2023, 137(1), 1047-1064. https://doi.org/10.32604/cmes.2023.027226

Abstract

Federated learning is a distributed machine learning method that can solve the increasingly serious problem of data islands and user data privacy, as it allows training data to be kept locally and not shared with other users. It trains a global model by aggregating locally-computed models of clients rather than their raw data. However, the divergence of local models caused by data heterogeneity of different clients may lead to slow convergence of the global model. For this problem, we focus on the client selection with federated learning, which can affect the convergence performance of the global model with the selected local models. We propose FedChoice, a client selection method based on loss function optimization, to select appropriate local models to improve the convergence of the global model. It firstly sets selected probability for clients with the value of loss function, and the client with high loss will be set higher selected probability, which can make them more likely to participate in training. Then, it introduces a local control vector and a global control vector to predict the local gradient direction and global gradient direction, respectively, and calculates the gradient correction vector to correct the gradient direction to reduce the cumulative deviation of the local gradient caused by the Non-IID data. We make experiments to verify the validity of FedChoice on CIFAR-10, CINIC-10, MNIST, EMNITS, and FEMNIST datasets, and the results show that the convergence of FedChoice is significantly improved, compared with FedAvg, FedProx, and FedNova.

Graphical Abstract

A Client Selection Method Based on Loss Function Optimization for Federated Learning

Keywords


Cite This Article

APA Style
Zeng, Y., Teng, S., Xiang, T., Zhang, J., Mu, Y. et al. (2023). A client selection method based on loss function optimization for federated learning. Computer Modeling in Engineering & Sciences, 137(1), 1047-1064. https://doi.org/10.32604/cmes.2023.027226
Vancouver Style
Zeng Y, Teng S, Xiang T, Zhang J, Mu Y, Ren Y, et al. A client selection method based on loss function optimization for federated learning. Comput Model Eng Sci. 2023;137(1):1047-1064 https://doi.org/10.32604/cmes.2023.027226
IEEE Style
Y. Zeng et al., "A Client Selection Method Based on Loss Function Optimization for Federated Learning," Comput. Model. Eng. Sci., vol. 137, no. 1, pp. 1047-1064. 2023. https://doi.org/10.32604/cmes.2023.027226



cc This work is licensed under a Creative Commons Attribution 4.0 International License , which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
  • 845

    View

  • 543

    Download

  • 0

    Like

Share Link