Abstract
We consider the straggler problem in decentralized learning over a logical ring while preserving user data privacy. Especially, we extend the recently proposed framework of differential privacy (DP) amplification by decentralization by Cyffers and Bellet to include overall training latency - comprising both computation and communication latency. Analytical results on both the convergence speed and the DP level are derived for both a skipping scheme (which ignores the stragglers after a timeout) and a baseline scheme that waits for each node to finish before the training continues. A trade-off between overall training latency, accuracy, and privacy, parameterized by the timeout of the skipping scheme, is identified and empirically validated for logistic regression on a real-world dataset and for image classification using the MNIST and CIFAR-10 datasets.
Original language | English (US) |
---|---|
Pages (from-to) | 407-423 |
Number of pages | 17 |
Journal | IEEE Journal on Selected Areas in Information Theory |
Volume | 5 |
DOIs | |
State | Published - 2024 |
All Science Journal Classification (ASJC) codes
- Computer Networks and Communications
- Media Technology
- Artificial Intelligence
- Applied Mathematics
Keywords
- Decentralized learning
- differential privacy
- gradient descent
- privacy amplification
- straggler mitigation
- training latency