TY - GEN
T1 - Decentralized Federated Learning via SGD over Wireless D2D Networks
AU - Xing, Hong
AU - Simeone, Osvaldo
AU - Bi, Suzhi
N1 - Publisher Copyright:
© 2020 IEEE.
PY - 2020/5
Y1 - 2020/5
N2 - Federated Learning (FL), an emerging paradigm for fast intelligent acquisition at the network edge, enables joint training of a machine learning model over distributed data sets and computing resources with limited disclosure of local data. Communication is a critical enabler of large-scale FL due to significant amount of model information exchanged among edge devices. In this paper, we consider a network of wireless devices sharing a common fading wireless channel for the deployment of FL. Each device holds a generally distinct training set, and communication typically takes place in a Device-To-Device (D2D) manner. In the ideal case in which all devices within communication range can communicate simultaneously and noiselessly, a standard protocol that is guaranteed to converge to an optimal solution of the global empirical risk minimization problem under convexity and connectivity assumptions is Decentralized Stochastic Gradient Descent (DSGD). DSGD integrates local SGD steps with periodic consensus averages that require communication between neighboring devices. In this paper, wireless protocols are proposed that implement DSGD by accounting for the presence of path loss, fading, blockages, and mutual interference. The proposed protocols are based on graph coloring for scheduling and on both digital and analog transmission strategies at the physical layer, with the latter leveraging over-The-Air computing via sparsity-based recovery.
AB - Federated Learning (FL), an emerging paradigm for fast intelligent acquisition at the network edge, enables joint training of a machine learning model over distributed data sets and computing resources with limited disclosure of local data. Communication is a critical enabler of large-scale FL due to significant amount of model information exchanged among edge devices. In this paper, we consider a network of wireless devices sharing a common fading wireless channel for the deployment of FL. Each device holds a generally distinct training set, and communication typically takes place in a Device-To-Device (D2D) manner. In the ideal case in which all devices within communication range can communicate simultaneously and noiselessly, a standard protocol that is guaranteed to converge to an optimal solution of the global empirical risk minimization problem under convexity and connectivity assumptions is Decentralized Stochastic Gradient Descent (DSGD). DSGD integrates local SGD steps with periodic consensus averages that require communication between neighboring devices. In this paper, wireless protocols are proposed that implement DSGD by accounting for the presence of path loss, fading, blockages, and mutual interference. The proposed protocols are based on graph coloring for scheduling and on both digital and analog transmission strategies at the physical layer, with the latter leveraging over-The-Air computing via sparsity-based recovery.
KW - D2D networks
KW - Federated learning
KW - distributed learning
KW - over-The-Air computing
UR - http://www.scopus.com/inward/record.url?scp=85090397425&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85090397425&partnerID=8YFLogxK
U2 - 10.1109/SPAWC48557.2020.9154332
DO - 10.1109/SPAWC48557.2020.9154332
M3 - Conference contribution
AN - SCOPUS:85090397425
T3 - IEEE Workshop on Signal Processing Advances in Wireless Communications, SPAWC
BT - 2020 IEEE 21st International Workshop on Signal Processing Advances in Wireless Communications, SPAWC 2020
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 21st IEEE International Workshop on Signal Processing Advances in Wireless Communications, SPAWC 2020
Y2 - 26 May 2020 through 29 May 2020
ER -