Non-parallel support vector classifiers with different loss functions

Siamak Mehrkanoon*, Xiaolin Huang, Johan A.K. Suykens

*Corresponding author for this work

Research output: Contribution to journalArticleAcademicpeer-review

Abstract

This paper introduces a general framework of non-parallel support vector machines, which involves a regularization term, a scatter loss and a misclassification loss. When dealing with binary problems, the framework with proper losses covers some existing non-parallel classifiers, such as multisurface proximal support vector machine via generalized eigenvalues, twin support vector machines, and its least squares version. The possibility of incorporating different existing scatter and misclassification loss functions into the general framework is discussed. Moreover, in contrast with the mentioned methods, which applies kernel-generated surface, we directly apply the kernel trick in the dual and then obtain nonparametric models. Therefore, one does not need to formulate two different primal problems for the linear and nonlinear kernel respectively. In addition, experimental results are given to illustrate the performance of different loss functions.

Original languageEnglish
Pages (from-to)294-301
Number of pages8
JournalNeurocomputing
Volume143
DOIs
Publication statusPublished - 2 Nov 2014

Bibliographical note

Funding Information:
EU: The research leading to these results has received funding from the European Research Council under the European Union׳s Seventh Framework Programme (FP7/2007-2013) / ERC AdG A-DATADRIVE-B (290923). This paper reflects only the authors׳ views, the Union is not liable for any use that may be made of the contained information. Research Council KUL: GOA/10/09 MaNet, CoE PFV/10/002 (OPTEC), BIL12/11T; PhD/Postdoc grants Flemish Government: FWO: projects: G.0377.12 (Structured systems), G.088114N (Tensor based data similarity); PhD/Postdoc grants IWT: projects: SBO POM (100031); PhD/Postdoc grants iMinds Medical Information Technologies SBO 2014 Belgian Federal Science Policy Office: IUAP P7/19 (DYSCO, Dynamical systems, control and optimization, 2012-2017). Johan Suykens is a professor at the KU Leuven, Belgium.

Funding

EU: The research leading to these results has received funding from the European Research Council under the European Union׳s Seventh Framework Programme (FP7/2007-2013) / ERC AdG A-DATADRIVE-B (290923). This paper reflects only the authors׳ views, the Union is not liable for any use that may be made of the contained information. Research Council KUL: GOA/10/09 MaNet, CoE PFV/10/002 (OPTEC), BIL12/11T; PhD/Postdoc grants Flemish Government: FWO: projects: G.0377.12 (Structured systems), G.088114N (Tensor based data similarity); PhD/Postdoc grants IWT: projects: SBO POM (100031); PhD/Postdoc grants iMinds Medical Information Technologies SBO 2014 Belgian Federal Science Policy Office: IUAP P7/19 (DYSCO, Dynamical systems, control and optimization, 2012-2017). Johan Suykens is a professor at the KU Leuven, Belgium.

Keywords

  • Hinge loss
  • Kernel trick
  • Least squares loss
  • Non-parallel classifiers
  • Pinball loss

Fingerprint

Dive into the research topics of 'Non-parallel support vector classifiers with different loss functions'. Together they form a unique fingerprint.

Cite this