Open Access
March 2011 Data augmentation for support vector machines
Nicholas G. Polson, Steven L. Scott
Bayesian Anal. 6(1): 1-23 (March 2011). DOI: 10.1214/11-BA601

Abstract

This paper presents a latent variable representation of regularized support vector machines (SVM's) that enables EM, ECME or MCMC algorithms to provide parameter estimates. We verify our representation by demonstrating that minimizing the SVM optimality criterion together with the parameter regularization penalty is equivalent to finding the mode of a mean-variance mixture of normals pseudo-posterior distribution. The latent variables in the mixture representation lead to EM and ECME point estimates of SVM parameters, as well as MCMC algorithms based on Gibbs sampling that can bring Bayesian tools for Gaussian linear models to bear on SVM's. We show how to implement SVM's with spike-and-slab priors and run them against data from a standard spam filtering data set.

Citation

Download Citation

Nicholas G. Polson. Steven L. Scott. "Data augmentation for support vector machines." Bayesian Anal. 6 (1) 1 - 23, March 2011. https://doi.org/10.1214/11-BA601

Information

Published: March 2011
First available in Project Euclid: 13 June 2012

zbMATH: 1330.62259
MathSciNet: MR2781803
Digital Object Identifier: 10.1214/11-BA601

Subjects:
Primary: 62H30
Secondary: 62C10 , 62J07 , 65C05

Keywords: $L^\alpha$-norm , Bayesian inference , ECME , EM , Lasso , MCMC , MCMC , regularization

Rights: Copyright © 2011 International Society for Bayesian Analysis

Vol.6 • No. 1 • March 2011
Back to Top