Dropout training for SVMs with data augmentation
From MaRDI portal
Publication:1713848
DOI10.1007/S11704-018-7314-7zbMATH Open1405.68280arXiv1508.02268OpenAlexW1922659888MaRDI QIDQ1713848FDOQ1713848
Authors: Ning Chen, J. Zhu, Jianfei Chen, Ting Chen
Publication date: 30 January 2019
Published in: Frontiers of Computer Science (Search for Journal in Brave)
Abstract: Dropout and other feature noising schemes have shown promising results in controlling over-fitting by artificially corrupting the training data. Though extensive theoretical and empirical studies have been performed for generalized linear models, little work has been done for support vector machines (SVMs), one of the most successful approaches for supervised learning. This paper presents dropout training for both linear SVMs and the nonlinear extension with latent representation learning. For linear SVMs, to deal with the intractable expectation of the non-smooth hinge loss under corrupting distributions, we develop an iteratively re-weighted least square (IRLS) algorithm by exploring data augmentation techniques. Our algorithm iteratively minimizes the expectation of a re-weighted least square problem, where the re-weights are analytically updated. For nonlinear latent SVMs, we consider learning one layer of latent representations in SVMs and extend the data augmentation technique in conjunction with first-order Taylor-expansion to deal with the intractable expected non-smooth hinge loss and the nonlinearity of latent representations. Finally, we apply the similar data augmentation ideas to develop a new IRLS algorithm for the expected logistic loss under corrupting distributions, and we further develop a non-linear extension of logistic regression by incorporating one layer of latent representations. Our algorithms offer insights on the connection and difference between the hinge loss and logistic loss in dropout training. Empirical results on several real datasets demonstrate the effectiveness of dropout training on significantly boosting the classification accuracy of both linear and nonlinear SVMs. In addition, the nonlinear SVMs further improve the prediction performance on several image datasets.
Full work available at URL: https://arxiv.org/abs/1508.02268
Recommendations
Learning and adaptive systems in artificial intelligence (68T05) Generalized linear models (logistic models) (62J12)
Cites Work
- Title not available (Why is that?)
- Title not available (Why is that?)
- In defense of one-vs-all classification
- Title not available (Why is that?)
- Bayesian Inference for Logistic Models Using Pólya–Gamma Latent Variables
- On the limited memory BFGS method for large scale optimization
- Data augmentation for support vector machines
- Learning to classify with missing and corrupted features
- Are Loss Functions All the Same?
- Title not available (Why is that?)
- On the inductive bias of dropout
- The dropout learning algorithm
- Title not available (Why is that?)
Cited In (1)
Uses Software
This page was built for publication: Dropout training for SVMs with data augmentation
Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q1713848)