TY - JOUR
T1 - Differentially Private Federated Learning With Stragglers' Delays in Cross-Silo Settings: An Online Mirror Descent Approach
AU - Odeyomi, Olusola
AU - Tankard, Earl
AU - Rawat, Danda
PY - 2024/2/1
Y1 - 2024/2/1
N2 - Federated learning is a privacy-preserving machine learning paradigm to protect the data of clients against privacy breaches. A lot of work on federated learning considers the cross-device setting where the number of clients is large and the data sample size of each client is low. However, this work focuses on cross-silo settings, where clients are few and have large sample sizes. We consider a fully decentralized setting where clients communicate with their immediate time-varying neighbors without the need for a central aggregator prone to congestion and a single point of failure. Our goal is to address stragglers' delays in cross-silo settings. Existing algorithms designed to overcome stragglers' delays work with fixed data distributions. They cannot work in real-time settings, such as wireless communication, characterized by time-varying data distributions. Therefore, this paper proposes two online learning algorithms that work with time-varying data and address stragglers' delays while guaranteeing differential privacy, strong convergence, and communication efficiency. Using the mirror descent technique, the first proposed algorithm addresses the case where the loss gradient is easily computed while the second proposed algorithm addresses the case where the loss gradient is difficult to compute. Simulation results show the performance of the proposed algorithms.
AB - Federated learning is a privacy-preserving machine learning paradigm to protect the data of clients against privacy breaches. A lot of work on federated learning considers the cross-device setting where the number of clients is large and the data sample size of each client is low. However, this work focuses on cross-silo settings, where clients are few and have large sample sizes. We consider a fully decentralized setting where clients communicate with their immediate time-varying neighbors without the need for a central aggregator prone to congestion and a single point of failure. Our goal is to address stragglers' delays in cross-silo settings. Existing algorithms designed to overcome stragglers' delays work with fixed data distributions. They cannot work in real-time settings, such as wireless communication, characterized by time-varying data distributions. Therefore, this paper proposes two online learning algorithms that work with time-varying data and address stragglers' delays while guaranteeing differential privacy, strong convergence, and communication efficiency. Using the mirror descent technique, the first proposed algorithm addresses the case where the loss gradient is easily computed while the second proposed algorithm addresses the case where the loss gradient is difficult to compute. Simulation results show the performance of the proposed algorithms.
KW - Differential privacy
KW - federated learning
KW - mirror descent
KW - online learning
KW - regret
UR - https://www.scopus.com/inward/record.uri?partnerID=HzOxMe3b&scp=85174803974&origin=inward
UR - https://www.scopus.com/inward/citedby.uri?partnerID=HzOxMe3b&scp=85174803974&origin=inward
U2 - 10.1109/TCCN.2023.3325815
DO - 10.1109/TCCN.2023.3325815
M3 - Article
SN - 2332-7731
VL - 10
SP - 308
EP - 321
JO - IEEE Transactions on Cognitive Communications and Networking
JF - IEEE Transactions on Cognitive Communications and Networking
IS - 1
ER -