TY - GEN
T1 - Multi-stage convex relaxation for learning with sparse regularization
AU - Zhang, Tong
PY - 2009
Y1 - 2009
N2 - We study learning formulations with non-convex regularizaton that are natural for sparse linear models. There are two approaches to this problem: • Heuristic methods such as gradient descent that only find a local minimum. A drawback of this approach is the lack of theoretical guarantee showing that the local minimum gives a good solution. • Convex relaxation such as L1-regularization that solves the problem under some conditions. However it often leads to sub-optimal sparsity in reality. This paper tries to remedy the above gap between theory and practice. In particular, we investigate a multi-stage convex relaxation scheme for solving problems with non-convex regularization. Theoretically, we analyze the behavior of a resulting two-stage relaxation scheme for the capped-L1 regularization. Our performance bound shows that the procedure is superior to the standard L 1 convex relaxation for learning sparse targets. Experiments confirm the effectiveness of this method on some simulation and real data.
AB - We study learning formulations with non-convex regularizaton that are natural for sparse linear models. There are two approaches to this problem: • Heuristic methods such as gradient descent that only find a local minimum. A drawback of this approach is the lack of theoretical guarantee showing that the local minimum gives a good solution. • Convex relaxation such as L1-regularization that solves the problem under some conditions. However it often leads to sub-optimal sparsity in reality. This paper tries to remedy the above gap between theory and practice. In particular, we investigate a multi-stage convex relaxation scheme for solving problems with non-convex regularization. Theoretically, we analyze the behavior of a resulting two-stage relaxation scheme for the capped-L1 regularization. Our performance bound shows that the procedure is superior to the standard L 1 convex relaxation for learning sparse targets. Experiments confirm the effectiveness of this method on some simulation and real data.
UR - http://www.scopus.com/inward/record.url?scp=84863420367&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=84863420367&partnerID=8YFLogxK
M3 - Conference contribution
AN - SCOPUS:84863420367
SN - 9781605609492
T3 - Advances in Neural Information Processing Systems 21 - Proceedings of the 2008 Conference
SP - 1929
EP - 1936
BT - Advances in Neural Information Processing Systems 21 - Proceedings of the 2008 Conference
PB - Neural Information Processing Systems
T2 - 22nd Annual Conference on Neural Information Processing Systems, NIPS 2008
Y2 - 8 December 2008 through 11 December 2008
ER -