Abstract:
|
The support vector machine (SVM) is a powerful binary classification tool widely used in medical and biological science. Since redundant features will severely affect the performance of SVM, penalized regression methods, especially those based on the lasso of Tibshirani (1996), have been utilized on its objective function for simultaneous variable selection and coefficient estimation. In this work, we investigate a general class of penalized SVMs under high dimensional settings. Properties such as variable consistency, estimator error bounds have been studied, and are achieved under mild conditions. Our analysis reveals that the method achieves nearly oracle performance, i.e. with large probability, the l2 norm of the estimation error is of order O(?klogp/n) under lasso penalty. The proposed algorithm can identify the oracle solution among potential local minimizers under some sufficient conditions. Simulation studies are provided as support for penalized support vector machine including other non-convex penalties.
|
ASA Meetings Department
732 North Washington Street, Alexandria, VA 22314
(703) 684-1221 • meetings@amstat.org
Copyright © American Statistical Association.