TY - JOUR
T1 - Information-based complexity, feedback and dynamics in convex programming
AU - Raginsky, Maxim
AU - Rakhlin, Alexander
N1 - Funding Information:
Manuscript received October 11, 2010; revised March 28, 2011; accepted March 28, 2011. Date of current version October 07, 2011. M. Raginsky was supported in part by the NSF under grant CCF-1017564. A. Rakhlin was supported by the NSF CAREER award DMS-0954737. A preliminary version of this work was presented at the 47th Annual Allerton Conference on Communication, Control, and Computing, Monticello, IL, September/October 2009.
PY - 2011/10
Y1 - 2011/10
N2 - We study the intrinsic limitations of sequential convex optimization through the lens of feedback information theory. In the oracle model of optimization, an algorithm queries an oracle for noisy information about the unknown objective function and the goal is to (approximately) minimize every function in a given class using as few queries as possible. We show that, in order for a function to be optimized, the algorithm must be able to accumulate enough information about the objective. This, in turn, puts limits on the speed of optimization under specific assumptions on the oracle and the type of feedback. Our techniques are akin to the ones used in statistical literature to obtain minimax lower bounds on the risks of estimation procedures; the notable difference is that, unlike in the case of i.i.d. data, a sequential optimization algorithm can gather observations in a controlled manner, so that the amount of information at each step is allowed to change in time. In particular, we show that optimization algorithms often obey the law of diminishing returns: the signal-to-noise ratio drops as the optimization algorithm approaches the optimum. To underscore the generality of the tools, we use our approach to derive fundamental lower bounds for a certain active learning problem. Overall, the present work connects the intuitive notions of information in optimization, experimental design, estimation, and active learning to the quantitative notion of Shannon information.
AB - We study the intrinsic limitations of sequential convex optimization through the lens of feedback information theory. In the oracle model of optimization, an algorithm queries an oracle for noisy information about the unknown objective function and the goal is to (approximately) minimize every function in a given class using as few queries as possible. We show that, in order for a function to be optimized, the algorithm must be able to accumulate enough information about the objective. This, in turn, puts limits on the speed of optimization under specific assumptions on the oracle and the type of feedback. Our techniques are akin to the ones used in statistical literature to obtain minimax lower bounds on the risks of estimation procedures; the notable difference is that, unlike in the case of i.i.d. data, a sequential optimization algorithm can gather observations in a controlled manner, so that the amount of information at each step is allowed to change in time. In particular, we show that optimization algorithms often obey the law of diminishing returns: the signal-to-noise ratio drops as the optimization algorithm approaches the optimum. To underscore the generality of the tools, we use our approach to derive fundamental lower bounds for a certain active learning problem. Overall, the present work connects the intuitive notions of information in optimization, experimental design, estimation, and active learning to the quantitative notion of Shannon information.
KW - Convex optimization
KW - Fano's inequality
KW - feedback information theory
KW - hypothesis testing with controlled observations
KW - information-based
KW - information-theoretic converse
KW - minimax lower bounds
KW - sequential optimization algorithms
KW - statistical estimation complexity
UR - http://www.scopus.com/inward/record.url?scp=80053997013&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=80053997013&partnerID=8YFLogxK
U2 - 10.1109/TIT.2011.2154375
DO - 10.1109/TIT.2011.2154375
M3 - Article
AN - SCOPUS:80053997013
SN - 0018-9448
VL - 57
SP - 7036
EP - 7056
JO - IRE Professional Group on Information Theory
JF - IRE Professional Group on Information Theory
IS - 10
M1 - 5766746
ER -