Hossein Moosaei and Milan Hladík. Least squares approach to K-SVCR multi-class classification with its applications. Ann. Math. Artif. Intell., 90:873–892, 2022.
[PDF] [gzipped postscript] [postscript] [HTML]
The support vector classification-regression machine for K-class classification (K-SVCR) is a novel multi-class classification method based on the 1-versus-1-versus-rest structure. In this paper, we propose a least squares version of K-SVCR named LSK-SVCR. Similarly to the K-SVCR algorithm, this method assesses all the training data into a 1-versus-1-versus-rest structure, so that the algorithm generates ternary outputs −1,0,+1. In LSK-SVCR, the solution of the primal problem is computed by solving only one system of linear equations instead of solving the dual problem, which is a convex quadratic programming problem in K-SVCR. Experimental results on several benchmark, MC-NDC, and handwritten digit recognition data sets show that not only does the LSK-SVCR have better performance in the aspects of classification accuracy to that of K-SVCR and Twin-KSVC algorithms but also has remarkably higher learning speed.
@article{MooHla2022a, author = "Hossein Moosaei and Milan Hlad\'{\i}k", title = "Least squares approach to {K-SVCR} multi-class classification with its applications", journal = "Ann. Math. Artif. Intell.", fjournal = "Annals of Mathematics and Artificial Intelligence", volume = "90", pages = "873-892", year = "2022", doi = "10.1007/s10472-021-09747-1", issn = "1573-7470", url = "https://link.springer.com/article/10.1007/s10472-021-09747-1", bib2html_dl_html = "https://doi.org/10.1007/s10472-021-09747-1", bib2html_dl_pdf = "https://rdcu.be/cT6ar", abstract = "The support vector classification-regression machine for K-class classification (K-SVCR) is a novel multi-class classification method based on the 1-versus-1-versus-rest structure. In this paper, we propose a least squares version of K-SVCR named LSK-SVCR. Similarly to the K-SVCR algorithm, this method assesses all the training data into a 1-versus-1-versus-rest structure, so that the algorithm generates ternary outputs {−1,0,+1}. In LSK-SVCR, the solution of the primal problem is computed by solving only one system of linear equations instead of solving the dual problem, which is a convex quadratic programming problem in K-SVCR. Experimental results on several benchmark, MC-NDC, and handwritten digit recognition data sets show that not only does the LSK-SVCR have better performance in the aspects of classification accuracy to that of K-SVCR and Twin-KSVC algorithms but also has remarkably higher learning speed.", keywords = "Support vector machine; Twin-KSVC; K-SVCR; Multi-class classification; Least squares", }
Generated by bib2html.pl (written by Patrick Riley ) on Wed Oct 23, 2024 08:16:44