Milan Hladík's Publications:

Sparse solution of least-squares twin multi-class support vector machine using $\ell_0$ and $\ell_p$-norm for classification and feature selection

Hossein Moosaei and Milan Hladík. Sparse solution of least-squares twin multi-class support vector machine using l0 and lp-norm for classification and feature selection. Neural Netw., 166:471–486, September 2023.

Download

[PDF] [gzipped postscript] [postscript] [HTML] 

Abstract

In the realm of multi-class classification, the twin K-class support vector classification (Twin-KSVC) generates ternary outputs -1,0,+1 by evaluating all training data in a `1-versus-1-versus-rest` structure. Recently, inspired by the least-squares version of Twin-KSVC and Twin-KSVC, a new multi-class classifier called improvements on least-squares twin multi-class classification support vector machine (ILSTKSVC) has been proposed. In this method, the concept of structural risk minimization is achieved by incorporating a regularization term in addition to the minimization of empirical risk. Twin-KSVC and its improvements have an influence on classification accuracy. Another aspect influencing classification accuracy is feature selection, which is a critical stage in machine learning, especially when working with high-dimensional datasets. However, most prior studies have not addressed this crucial aspect. In this study, motivated by ILSTKSVC and the cardinality-constrained optimization problem, we propose $\ell_p$-norm least-squares twin multi-class support vector machine (PLSTKSVC) with 0 < p < 1 to perform classification and feature selection at the same time. The technique employed to solve the optimization problems associated with PLSTKSVC is user-friendly, as it involves solving systems of linear equations to obtain an approximate solution for the proposed model. Under certain assumptions, we investigate the properties of the optimum solutions to the related optimization problems. Several real-world datasets were tested using the suggested method. According to the results of our experiments, the proposed method outperforms all current strategies in most datasets in terms of classification accuracy while also reducing the number of features.

BibTeX

@article{MooHla2023c,
 author = "Hossein Moosaei and Milan Hlad\'{\i}k",
 title = "Sparse solution of least-squares  twin multi-class support vector machine using $\ell_{0}$ and $\ell_p$-norm for classification and feature selection",
 webtitle = "Sparse solution of least-squares  twin multi-class support vector machine using l<sub>0</sub> and l<sub>p</sub>-norm for classification and feature selection",
 journal = "Neural Netw.",
 fjournal = "Neural Networks",
 volume = "166",
 month = "September",
 pages = "471-486",
 year = "2023",
 doi = "10.1016/j.neunet.2023.07.039",
 issn = "0893-6080",
 url = "https://www.sciencedirect.com/science/article/pii/S0893608023003982",
 bib2html_dl_html = "https://doi.org/10.1016/j.neunet.2023.07.039",
 bib2html_dl_pdf = "https://doi.org/10.1016/j.neunet.2023.07.039",
 abstract = "In the realm of multi-class classification, the twin K-class support vector classification (Twin-KSVC) generates ternary outputs {-1,0,+1} by evaluating all training data in a `1-versus-1-versus-rest` structure.  Recently, inspired by the least-squares version of Twin-KSVC and Twin-KSVC, a new multi-class classifier called improvements on least-squares twin multi-class classification support vector machine (ILSTKSVC) has been proposed. In this method, the concept of structural risk minimization is achieved by incorporating a regularization term in addition to the minimization of empirical risk. Twin-KSVC and its improvements have an influence on classification accuracy. Another aspect influencing classification accuracy is feature selection, which is a critical stage in machine learning, especially when working with high-dimensional datasets. However, most prior studies have not addressed this crucial aspect. In this study, motivated by ILSTKSVC and the cardinality-constrained optimization problem, we propose $\ell_p$-norm least-squares twin multi-class support vector machine (PLSTKSVC) with 0 < p < 1 to perform classification and feature selection at the same time. The technique employed to solve the optimization problems associated with PLSTKSVC is user-friendly, as it involves solving systems of linear equations to obtain an approximate solution for the proposed model. Under certain assumptions, we investigate the properties of the optimum solutions to the related optimization problems. Several real-world datasets were tested using the suggested method. According to the results of our experiments, the proposed method outperforms all current strategies in most datasets in terms of classification accuracy while also reducing the number of features.",
 keywords = "Multi-class classification; Twin k-class support vector classification; Least-squares; Cardinality-constrained optimization problem; Lp-norm; Feature selection",
}

Generated by bib2html.pl (written by Patrick Riley ) on Mon Apr 15, 2024 08:26:42