Algorithms for Sparse Support Vector Machines

From MaRDI portal
Publication:6180740

DOI10.1080/10618600.2022.2146697arXiv2110.07691MaRDI QIDQ6180740FDOQ6180740


Authors: Kenneth Lange Edit this on Wikidata


Publication date: 22 January 2024

Published in: Journal of Computational and Graphical Statistics (Search for Journal in Brave)

Abstract: Many problems in classification involve huge numbers of irrelevant features. Model selection reveals the crucial features, reduces the dimensionality of feature space, and improves model interpretation. In the support vector machine literature, model selection is achieved by ell1 penalties. These convex relaxations seriously bias parameter estimates toward 0 and tend to admit too many irrelevant features. The current paper presents an alternative that replaces penalties by sparse-set constraints. Penalties still appear, but serve a different purpose. The proximal distance principle takes a loss function and adds the penalty capturing the squared Euclidean distance of the parameter vector to the sparsity set Sk where at most k components of are nonzero. If represents the minimum of the objective , then tends to the constrained minimum of over Sk as ho tends to infty. We derive two closely related algorithms to carry out this strategy. Our simulated and real examples vividly demonstrate how the algorithms achieve much better sparsity without loss of classification power.


Full work available at URL: https://arxiv.org/abs/2110.07691







Cites Work






This page was built for publication: Algorithms for Sparse Support Vector Machines

Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q6180740)