On the consistency of feature selection using greedy least squares regression

Research output: Contribution to journalArticlepeer-review

Abstract

This paper studies the feature selection problem using a greedy least squares regression algorithm. We show that under a certain irrepresentable condition on the design matrix (but independent of the sparse target), the greedy algorithm can select features consistently when the sample size approaches infinity. The condition is identical to a corresponding condition for Lasso. Moreover, under a sparse eigenvalue condition, the greedy algorithm can reliably identify features as long as each nonzero coefficient is larger than a constant times the noise level. In comparison, Lasso may require the coefficients to be larger than 0(√s) times the noise level in the worst case, where s is the number of nonzero coefficients.

Original languageEnglish (US)
Pages (from-to)555-568
Number of pages14
JournalJournal of Machine Learning Research
Volume10
StatePublished - Jan 2009
Externally publishedYes

Keywords

  • Feature selection
  • Greedy algorithm
  • Sparsity

ASJC Scopus subject areas

  • Software
  • Control and Systems Engineering
  • Statistics and Probability
  • Artificial Intelligence

Fingerprint

Dive into the research topics of 'On the consistency of feature selection using greedy least squares regression'. Together they form a unique fingerprint.

Cite this