Support vector machines with two support vectors
In this article we present a new class of support vector machines for binary classification task. Our support vector machines are constructed using only two support vectors and have very low Vapnik-Chervonenkis dimension, so they generalize well. Geometrically, our approach is based on searching of a proper pair of observations from different classes of explained variable. Once this pair is found the discriminant hyperplane becomes orthogonal to the line connecting these observations. This method deals well with data sets with large number of features and small number of observations like gene expression data. We illustrate the performance of our classification method using gene expression data and show that it is superior to other classifiers especially to diagonal linear discriminant analysis and k-nearest neighbor which achieved the lowest error rate in the previous studies of tumor classification.
|Date of creation:||01 May 2009|
|Date of revision:|
|Contact details of provider:|| Postal: |
Phone: + (48)(22) 49 12 51
Fax: + (48)(22) 49 53 12
Web page: http://www.sgh.waw.pl/instytuty/zes
More information through EDIRC
When requesting a correction, please mention this item's handle: RePEc:wse:wpaper:35. See general information about how to correct material in RePEc.
For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: (Marcin Owczarczuk)
If references are entirely missing, you can add them using this form.