Latency-Efficient Wireless Federated Learning With Sparsification and Quantization for Heterogeneous Devices
Recently, federated learning (FL) has attracted much attention as a promising decentralized machine learning method that provides privacy and low latency. However, the communication bottleneck is still a problem that needs to be solved to effectively deploy FL on wireless networks. In this article,...
Saved in:
| Published in | IEEE internet of things journal Vol. 12; no. 1; pp. 488 - 501 |
|---|---|
| Main Authors | , , , |
| Format | Journal Article |
| Language | English |
| Published |
IEEE
01.01.2025
|
| Subjects | |
| Online Access | Get full text |
| ISSN | 2327-4662 2327-4662 |
| DOI | 10.1109/JIOT.2024.3462722 |
Cover
| Summary: | Recently, federated learning (FL) has attracted much attention as a promising decentralized machine learning method that provides privacy and low latency. However, the communication bottleneck is still a problem that needs to be solved to effectively deploy FL on wireless networks. In this article, we aim to minimize the total convergence time of FL by sparsifying and quantizing local model parameters before uplink transmission. More specifically, we first present the convergence analysis of the FL algorithm with random sparsification and quantization, revealing the impact of compression error on the convergence speed. Then, we jointly optimize the computation, communication resources and the number of quantization bits, sparsity to minimize the total convergence time, subject to the energy and compression error requirements derived from the convergence analysis. By simulating the impact of different compression errors on model accuracy, we reveal that the low-precision updates do not inherently yield a better balance between efficiency and accuracy than the high-precision updates. Furthermore, compared with the equal resource allocation schemes and the unilateral compression optimization schemes on four different data distributions, the proposed scheme has faster convergence speed and less total convergence time. |
|---|---|
| ISSN: | 2327-4662 2327-4662 |
| DOI: | 10.1109/JIOT.2024.3462722 |