Grey Wolf Optimizer for Reducing Communication Cost of Federated Learning
2022 IEEE Global Communications Conference, GLOBECOM 2022 - Proceedings
Federated Learning (FL) is a type of Machine Learning (ML) technique in which only learned models are stored on a server to sustain data security. The approach does not gather server-side data but rather directly shares only the models from scattered clients. Due to the fact that clients of FL frequently have restricted connection bandwidth, it is necessary to optimize the communication between servers and clients. FL clients frequently interact through Wi-Fi and must operate in uncertain network situations. Nevertheless, the enormous number of weights transmitted and received by existing FL aggregation techniques dramatically degrade the accuracy in unstable network situations. We propose a federated GWO (FedGWO) algorithm to reduce data communications. The proposed approach improves the performance under unstable network conditions by transferring score principles rather than all client models' weights. We achieve a 13.55% average improvement in the global model's accuracy while decreasing the data capacity required for network communication. Moreover, we show that FedGWO achieves a 5% reduction in accuracy loss compared to FedAvg and Federated Particle Swarm Optimization (FedPSO) methods when tested on unstable networks.
Aggregation, Convolutional Neural Network (CNN), Deep Learning, Federated Learning, Grey Wolf Optimizer (GWO)
A. K. Abasi, M. Aloqaily and M. Guizani, "Grey Wolf Optimizer for Reducing Communication Cost of Federated Learning," GLOBECOM 2022 - 2022 IEEE Global Communications Conference, Rio de Janeiro, Brazil, 2022, pp. 1049-1154, doi: 10.1109/GLOBECOM48099.2022.10001681.