Wireless Quantized Federated Learning: A Joint Computation and Communication Design

Recently, federated learning (FL) has sparked widespread attention as a promising decentralized machine learning approach which provides privacy and low delay. However, communication bottleneck still constitutes an issue, that needs to be resolved for an efficient deployment of FL over wireless netw...

Full description

Saved in:
Bibliographic Details
Published inIEEE transactions on communications Vol. 71; no. 5; pp. 2756 - 2770
Main Authors Bouzinis, Pavlos S., Diamantoulakis, Panagiotis D., Karagiannidis, George K.
Format Journal Article
LanguageEnglish
Published New York IEEE 01.05.2023
The Institute of Electrical and Electronics Engineers, Inc. (IEEE)
Subjects
Online AccessGet full text
ISSN0090-6778
1558-0857
1558-0857
DOI10.1109/TCOMM.2023.3258485

Cover

More Information
Summary:Recently, federated learning (FL) has sparked widespread attention as a promising decentralized machine learning approach which provides privacy and low delay. However, communication bottleneck still constitutes an issue, that needs to be resolved for an efficient deployment of FL over wireless networks. In this paper, we aim to minimize the total convergence time of FL, by quantizing the local model parameters prior to uplink transmission. More specifically, the convergence analysis of the FL algorithm with stochastic quantization is firstly presented, which reveals the impact of the quantization error on the convergence rate. Following that, we jointly optimize the computing and communication resources as well as the number of quantization bits, in order to guarantee minimized convergence time, subject to energy and quantization error requirements. The impact of the quantization error on the convergence time is evaluated and the trade-off among model accuracy and timely execution is revealed. Moreover, the proposed method is shown to result in faster convergence compared with baseline schemes. Finally, useful insights for the selection of the quantization error tolerance are provided.
Bibliography:ObjectType-Article-1
SourceType-Scholarly Journals-1
ObjectType-Feature-2
content type line 14
ISSN:0090-6778
1558-0857
1558-0857
DOI:10.1109/TCOMM.2023.3258485