Abstract
Large margin classifiers such as support vector machines (SVM) have been applied successfully in various classification tasks. However, their performance may be significantly degraded in the presence of outliers. In this paper, we propose a robust SVM formulation which is shown to be less sensitive to outliers. The key idea is to employ an adaptively weighted hinge loss that explicitly incorporates outlier filtering in the SVM training, thus performing outlier filtering and classification simultaneously. The resulting robust SVM formulation is non-convex. We first relax it into a semi-definite programming which admits a global solution. To improve the efficiency, an iterative approach is developed. We have performed experiments using both synthetic and real-world data. Results show that the performance of the standard SVM degrades rapidly when more outliers are included, while the proposed robust SVM training is more stable in the presence of outliers.
Original language | English (US) |
---|---|
Pages (from-to) | 362-370 |
Number of pages | 9 |
Journal | Journal of Zhejiang University: Science C |
Volume | 12 |
Issue number | 5 |
DOIs | |
State | Published - May 2011 |
Keywords
- Multi-stage relaxation
- Outlier filter
- Semi-definite programming
- Support vector machines
ASJC Scopus subject areas
- Engineering(all)