Techniques for Selecting the Optimal Parameters of One-Class Support Vector Machine Classifier for Reduced Samples

Techniques for Selecting the Optimal Parameters of One-Class Support Vector Machine Classifier for Reduced Samples

Yasmine Guerbai, Youcef Chibani, Yassine Meraihi
Copyright: © 2022 |Pages: 15
DOI: 10.4018/IJAMC.290533
Article PDF Download
Open access articles are freely available for download

Abstract

Usually, the One-Class Support Vector Machine (OC-SVM) requires a large dataset for modeling effectively the target class independently to other classes. For finding the OC-SVM model, the available dataset is subdivided into two subsets namely training and validation, which are used for training and validating the optimal parameters. This approach is effective when a large dataset is available. However, when training samples are reduced, parameters of the OC-SVM are difficult to find in absence of the validation subset. Hence, this paper proposes various techniques for selecting the optimal parameters using only a training subset. The experimental evaluation conducted on several real-world benchmarks proves the effective use of the new selection parameter techniques for validating the model of OC-SVM classifiers versus the standard validation techniques
Article Preview
Top

1 Introduction

In statistics or data mining, a typical task is to learn a model from available data. The problem with evaluating such a model is that it may demonstrate adequate prediction capability on the training data, but might fail to predict future unseen data. Also, the validation protocol is the most procedure for estimating the generalization performance in this context.

Generally, cross-validation is a statistical method of evaluating and comparing the training model by dividing data into two subsets; one is used for training and the other is used for evaluating the model. In typical cross-validation, the training and validation sets must cross over in successive rounds such that each data point has a chance of being validated against. The basic form of cross-validation is k-fold cross-validation. Other forms of cross-validation are special cases of k-fold cross-validation or involve repeated rounds of k-fold cross-validation, first appeared by Mosteller and Turkey. (Mosteller & Tukey, 1968)

In 1970, both Stone (Stone, 1974) and Geisser (Geisser, 1975) employed cross-validation. Currently, cross-validation is widely accepted in machine learning as a standard procedure for performance evaluation. There are two possible goals in cross-validation. The first goal is to estimate the performance of the learned model from available data using one algorithm. The second goal is to compare the performance of two or more different algorithms and find out the best algorithm for the available data, or alternatively to compare the performance of two or more variants of the parameterized model.

Complete Article List

Search this Journal:
Reset
Volume 15: 1 Issue (2024)
Volume 14: 1 Issue (2023)
Volume 13: 4 Issues (2022): 2 Released, 2 Forthcoming
Volume 12: 4 Issues (2021)
Volume 11: 4 Issues (2020)
Volume 10: 4 Issues (2019)
Volume 9: 4 Issues (2018)
Volume 8: 4 Issues (2017)
Volume 7: 4 Issues (2016)
Volume 6: 4 Issues (2015)
Volume 5: 4 Issues (2014)
Volume 4: 4 Issues (2013)
Volume 3: 4 Issues (2012)
Volume 2: 4 Issues (2011)
Volume 1: 4 Issues (2010)
View Complete Journal Contents Listing