TY - JOUR
T1 - Projected stochastic primal-dual method for constrained online learning with kernels
AU - Koppel, Alec
AU - Zhang, Kaiqing
AU - Zhu, Hao
AU - Basar, Tamer
N1 - Manuscript received April 30, 2018; revised January 17, 2019 and March 6, 2019; accepted March 13, 2019. Date of publication April 1, 2019; date of current version April 12, 2019. The associate editor coordinating the review of this manuscript and approving it for publication was Dr. Dennis Wei. This work was supported in part by the Army Research Laboratory under the Cooperative Agreement W911NF-17-2-0196, in part by the National Science Foundation under the Award ECCS-1802319, and in part by the ASEE SMART. This paper was presented in part at the IEEE 57th Annual Conference on Decision and Control, Miami Beach, FL, USA, December 2018 [1]. (Alec Koppel and Kaiqing Zhang contributed equally to this work.) (Corresponding author: Alec Koppel.) A. Koppel is with the U.S. Army Research Laboratory, Adelphi, MD 20783 USA (e-mail:,[email protected]).
PY - 2019/5/15
Y1 - 2019/5/15
N2 - We consider the problem of stochastic optimization with nonlinear constraints, where the decision variable is not vector-valued but instead a function belonging to a reproducing Kernel Hilbert Space (RKHS). Currently, there exist solutions to only special cases of this problem. To solve this constrained problem with kernels, we first generalize the Representer Theorem to a class of saddle-point problems defined over RKHS. Then, we develop a primal-dual method which that executes alternating projected primal/dual stochastic gradient descent/ascent on the dual-augmented Lagrangian of the problem. The primal projection sets are low-dimensional subspaces of the ambient function space, which are greedily constructed using matching pursuit. By tuning the projection-induced error to the algorithm step-size, we are able to establish mean convergence in both primal objective sub-optimality and constraint violation, to respective O(√T) and O}(T3/4) neighborhoods. Here, T is the final iteration index and the constant step-size is chosen as 1/√T with 1/T approximation budget. Finally, we demonstrate experimentally the effectiveness of the proposed method for risk-aware supervised learning.
AB - We consider the problem of stochastic optimization with nonlinear constraints, where the decision variable is not vector-valued but instead a function belonging to a reproducing Kernel Hilbert Space (RKHS). Currently, there exist solutions to only special cases of this problem. To solve this constrained problem with kernels, we first generalize the Representer Theorem to a class of saddle-point problems defined over RKHS. Then, we develop a primal-dual method which that executes alternating projected primal/dual stochastic gradient descent/ascent on the dual-augmented Lagrangian of the problem. The primal projection sets are low-dimensional subspaces of the ambient function space, which are greedily constructed using matching pursuit. By tuning the projection-induced error to the algorithm step-size, we are able to establish mean convergence in both primal objective sub-optimality and constraint violation, to respective O(√T) and O}(T3/4) neighborhoods. Here, T is the final iteration index and the constant step-size is chosen as 1/√T with 1/T approximation budget. Finally, we demonstrate experimentally the effectiveness of the proposed method for risk-aware supervised learning.
KW - Machine learning
KW - optimization methods
KW - radial basis function networks
UR - http://www.scopus.com/inward/record.url?scp=85064539124&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85064539124&partnerID=8YFLogxK
U2 - 10.1109/TSP.2019.2907265
DO - 10.1109/TSP.2019.2907265
M3 - Article
AN - SCOPUS:85064539124
SN - 1053-587X
VL - 67
SP - 2528
EP - 2542
JO - IEEE Transactions on Signal Processing
JF - IEEE Transactions on Signal Processing
IS - 10
M1 - 8678800
ER -