TY - GEN
T1 - Distributed Optimization over Time-Varying Networks
T2 - 2022 American Control Conference, ACC 2022
AU - Reisizadeh, Hadi
AU - Touri, Behrouz
AU - Mohajer, Soheil
N1 - H. Reisizadeh (email: [email protected]) and S. Mohajer (email: [email protected]) are with the University of Minnesota, and B. Touri (email: [email protected]) is with the University of California San Diego. The work of H. Reisizadeh and S. Mohajer is supported in part by the National Science Foundation under Grants CCF-1749981.
PY - 2022
Y1 - 2022
N2 - The convergence of an error-feedback algorithm is studied for decentralized stochastic gradient descent (DSGD) algorithm with compressed information sharing over time-varying graphs. It is shown that for both strongly-convex and convex cost functions, despite of imperfect information sharing, the convergence rates match those with perfect information sharing. To do so, we show that for strongly-convex loss functions, with a proper choice of a step-size, the state of each node converges to the global optimizer at the rate of mathcal{O}left( {T{-1}} right). Similarly, for general convex cost functions, with a proper choice of step-size, we show that the value of loss function at a temporal average of each node's estimates converges to the optimal value at the rate of mathcal{O}left( {T{-1/2 + varepsilon }} right) for any ϵ > 0.
AB - The convergence of an error-feedback algorithm is studied for decentralized stochastic gradient descent (DSGD) algorithm with compressed information sharing over time-varying graphs. It is shown that for both strongly-convex and convex cost functions, despite of imperfect information sharing, the convergence rates match those with perfect information sharing. To do so, we show that for strongly-convex loss functions, with a proper choice of a step-size, the state of each node converges to the global optimizer at the rate of mathcal{O}left( {T{-1}} right). Similarly, for general convex cost functions, with a proper choice of step-size, we show that the value of loss function at a temporal average of each node's estimates converges to the optimal value at the rate of mathcal{O}left( {T{-1/2 + varepsilon }} right) for any ϵ > 0.
UR - http://www.scopus.com/inward/record.url?scp=85138491448&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85138491448&partnerID=8YFLogxK
U2 - 10.23919/ACC53348.2022.9867680
DO - 10.23919/ACC53348.2022.9867680
M3 - Conference contribution
AN - SCOPUS:85138491448
T3 - Proceedings of the American Control Conference
SP - 2791
EP - 2796
BT - 2022 American Control Conference, ACC 2022
PB - Institute of Electrical and Electronics Engineers Inc.
Y2 - 8 June 2022 through 10 June 2022
ER -