Sparse Algorithms Are Not Stable: A No-Free-Lunch Theorem

We consider two desired properties of learning algorithms: sparsity and algorithmic stability. Both properties are believed to lead to good generalization ability. We show that these two properties are fundamentally at odds with each other: A sparse algorithm cannot be stable and vice versa. Thus, o...

Full description

Saved in:
Bibliographic Details
Published inIEEE transactions on pattern analysis and machine intelligence Vol. 34; no. 1; pp. 187 - 193
Main Authors Huan Xu, Caramanis, C., Mannor, S.
Format Journal Article
LanguageEnglish
Published Los Alamitos, CA IEEE 01.01.2012
IEEE Computer Society
The Institute of Electrical and Electronics Engineers, Inc. (IEEE)
Subjects
Online AccessGet full text
ISSN0162-8828
1939-3539
2160-9292
2160-9292
1939-3539
DOI10.1109/TPAMI.2011.177

Cover

More Information
Summary:We consider two desired properties of learning algorithms: sparsity and algorithmic stability. Both properties are believed to lead to good generalization ability. We show that these two properties are fundamentally at odds with each other: A sparse algorithm cannot be stable and vice versa. Thus, one has to trade off sparsity and stability in designing a learning algorithm. In particular, our general result implies that ℓ 1 -regularized regression (Lasso) cannot be stable, while ℓ 2 -regularized regression is known to have strong stability properties and is therefore not sparse.
Bibliography:ObjectType-Article-1
SourceType-Scholarly Journals-1
ObjectType-Feature-2
content type line 14
content type line 23
ObjectType-Article-2
ObjectType-Feature-1
ISSN:0162-8828
1939-3539
2160-9292
2160-9292
1939-3539
DOI:10.1109/TPAMI.2011.177