TY - JOUR
T1 - Adaptive forward-backward greedy algorithm for learning sparse representations
AU - Zhang, Tong
N1 - Manuscript received October 16, 2008; revised October 05, 2010; accepted December 31, 2010. Date of current version June 22, 2011. The author was supported in part by the following grants: AFOSR-10097389, NSA-AMS 081024, NSF DMS-1007527, and NSF IIS-1016061. The author is with the Statistics Department, Rutgers University, Piscataway NJ 08854 USA (e-mail: [email protected]). Communicated by A. Krzyzak, Associate Editor for Pattern Recognition, Statistical Learning, and Inference. Digital Object Identifier 10.1109/TIT.2011.2146690
PY - 2011/7
Y1 - 2011/7
N2 - Given a large number of basis functions that can be potentially more than the number of samples, we consider the problem of learning a sparse target function that can be expressed as a linear combination of a small number of these basis functions. We are interested in two closely related themes: • feature selection, or identifying the basis functions with nonzero coefficients; • estimation accuracy, or reconstructing the target function from noisy observations. Two heuristics that are widely used in practice are forward and backward greedy algorithms. First, we show that neither idea is adequate. Second, we propose a novel combination that is based on the forward greedy algorithm but takes backward steps adaptively whenever beneficial. For least squares regression, we develop strong theoretical results for the new procedure showing that it can effectively solve this problem under some assumptions. Experimental results support our theory.
AB - Given a large number of basis functions that can be potentially more than the number of samples, we consider the problem of learning a sparse target function that can be expressed as a linear combination of a small number of these basis functions. We are interested in two closely related themes: • feature selection, or identifying the basis functions with nonzero coefficients; • estimation accuracy, or reconstructing the target function from noisy observations. Two heuristics that are widely used in practice are forward and backward greedy algorithms. First, we show that neither idea is adequate. Second, we propose a novel combination that is based on the forward greedy algorithm but takes backward steps adaptively whenever beneficial. For least squares regression, we develop strong theoretical results for the new procedure showing that it can effectively solve this problem under some assumptions. Experimental results support our theory.
KW - Estimation theory
KW - feature selection
KW - greedy algorithms
KW - sparse recovery
KW - statistical learning
UR - https://www.scopus.com/pages/publications/79959549699
UR - https://www.scopus.com/inward/citedby.url?scp=79959549699&partnerID=8YFLogxK
U2 - 10.1109/TIT.2011.2146690
DO - 10.1109/TIT.2011.2146690
M3 - Article
AN - SCOPUS:79959549699
SN - 0018-9448
VL - 57
SP - 4689
EP - 4708
JO - IEEE Transactions on Information Theory
JF - IEEE Transactions on Information Theory
IS - 7
M1 - 5895111
ER -