Next Article in Journal
Harmonic and Vibration Analysis of Dual Three-Phase PMSM with One-Phase Open Circuit Fault
Previous Article in Journal
Numerical Study on the Reinforcement Measures of Tunneling on Adjacent Piles
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Novel Twin Support Vector Machine with Generalized Pinball Loss Function for Pattern Classification

by
Wanida Panup
1,
Wachirapong Ratipapongton
2 and
Rabian Wangkeeree
1,3,*
1
Department of Mathematics, Faculty of Science, Naresuan University, Phitsanulok 65000, Thailand
2
Department of Mathematics, University of York, Heslington, York YO10 5DD, UK
3
Research Center for Academic Excellence in Mathematics, Naresuan University, Phitsanulok 65000, Thailand
*
Author to whom correspondence should be addressed.
Symmetry 2022, 14(2), 289; https://doi.org/10.3390/sym14020289
Submission received: 23 December 2021 / Revised: 9 January 2022 / Accepted: 18 January 2022 / Published: 31 January 2022
(This article belongs to the Topic Topology vs. Geometry in Data Analysis/Machine Learning)

Abstract

:
We introduce a novel twin support vector machine with the generalized pinball loss function (GPin-TSVM) for solving data classification problems that are less sensitive to noise and preserve the sparsity of the solution. In addition, we use a symmetric kernel trick to enlarge GPin-TSVM to nonlinear classification problems. The developed approach is tested on numerous UCI benchmark datasets, as well as synthetic datasets in the experiments. The comparisons demonstrate that our proposed algorithm outperforms existing classifiers in terms of accuracy. Furthermore, this employed approach in handwritten digit recognition applications is examined, and the automatic feature extractor employs a convolution neural network.

1. Introduction

Support vector machines (SVMs) have evolved as a potent paradigm for pattern classification and regression during the last decade [1,2,3,4,5]. The SVM has received substantial attention within a few years after its inception due to its vast application in a wide variety of fields [6,7,8,9,10,11,12]. The standard SVM determines the parallel hyperplanes with the maximum margin between two classes of samples by minimizing structural and empirical risks, as determined by the labeled training data. SVM solves a quadratic programming problem (QPP) using the dual problem to achieve an optimal solution. The standard SVMs also faces a major challenge: the computational complexity of SVM is approximately of order O ( m 3 ) , where m is a number of training samples. As a result, SVM is quite slow when dealing with large-scale problems [13,14].
For large-scale data learning, Catak [15] combines the ELM algorithm and the AdaBoosting method to overcome large-scale data sets. Moreover, to address the high-computational complexity of SVM, Jayadeva [16] suggested a novel machine learning method known as twin SVM (TSVM) to improve the computational complexity of SVM. For the standard TSVM, the main idea is to find two nonparallel proximal hyperplanes that are closer to one of the two classes while being at least one distance apart. TSVM solves two smaller QPPs, instead of solving a large one as in the classical SVM. Therefore, it makes the computational time of TSVM approximately four times faster than the standard SVM in theory. In binary classification problems, TSVM not only overcomes the challenges of training a classifier faster than a standard SVM, but it also deals with exemplar unbalance. As a result of its excellent performance, TSVM has become one of the most used procedures. TSVM has received increasing attention due to its wide application in various fields, such as text categorization [17], text recognition [18], software defects [19,20], scene classification [21], image recognition [22], speaker recognition [23,24], human action recognition [25], pancreatic cancer early detection [26], and so on. Moreover, TSVM has been widely researched and developed in recent years. There have been numerous variations proposed, such as twin parametric margin SVM (TPMSVM) [27], twin bounded SVM (TBSVM) [28], weighted Lagrangian TSVM (WLTSVM) [29], least squares TSVM (LSTSVM) [30,31,32], large scale TSVM [33], sparse pinball TSVM [34], and so on. Furthermore, TSVM is very useful when dealing with datasets that include a large number of data samples, whereas the standard SVM is ineffective.
Designing a robust machine learning approach, on the one hand, needs the employment of the appropriate loss function. Different margin-based loss functions have recently been employed in classification and regression problems, such as 0–1 loss, hinge loss, squared loss, and so on. The hinge loss controls the penalty on the training data points in standard SVM and TSVM. There are some problems with the model itself, such as the objective function in the primal problem is non-differential, has imbalanced class information, is sensitive to outliers, and is sensitive to feature noise. To address the non-differentiality of the objective function, a smooth SVM (SSVM) [35] has been proposed, where the SSVM creates and solves an unconstrained smooth support vector machine reformulation. When there were outliers, Wu and Liu [36] proposed the robust truncated hinge loss SVM (RSVM) to overcome this problem. For the imbalanced classification problem, Cao and Shen [37] demonstrated a re-sampling strategy that balances training data by combining oversampling and under-sampling. In [38], a powerful weighted multi-class least squares TSVM (WMLSTSVM) method for dealing with multi-class data categorization imbalances was proposed. In the presence of being sensitive to noise, Huang [39] proposed a SVM model to deal with noise sensitivity and instability in resampling, where the pinball loss function (Pin-SVM) is used. The outcome has good properties, such as being less sensitive to noise and related to the quantile distance. However, sparsity is impossible to attain using Pin-SVM. In order to maintain the sparsity, they also proposed an ϵ -insensitive zone for Pin-SVM. Although this approach improves the sparsity of Pin-SVM, its formulation necessitates the specification of the value of ϵ in advance, and hence a poor choice may have an impact on its performance. As a result of these advances, Rastogi [40] recently proposed the modified ( ϵ 1 , ϵ 2 ) -insensitive zone SVM, which is called the generalized pinball loss SVM. This generalized pinball loss for the SVM model incorporates previous loss functions that provide noise sensitivity, sparsity, and approximate stability. Nevertheless, compared with TSVM, the loss of the generalized pinball SVM is indeed required to solve a single large QPP, resulting in a higher computational complexity and inability to solve large scale problems. However, as far as we are aware, no articles dealing with the generalized pinball loss function in relation to the standard TSVM for classification problems have been published. As a result, the addition of the generalized pinball loss function to the standard TSVM for classification problems is worth investigating. Motivated by the above mentioned models, we introduce the standard TSVM with the generalized pinball loss function. In addition, the proposed objective function is optimized using the Lagrangian multiplier approach and the Karush–Kuhn–Tucker (KKT) optimality conditions [41]. Two smaller quadratic programming problems are solved (QPPs), and we can produce two nonparallel classification hyperplanes. Finally, thorough experiments were carried out to evaluate the proposed GPin-TSVM model performance. The following are the main contributions of the paper:
  • For pattern classification, we add a generalized pinball loss function to the standard TSVM, resulting in a better classifier model that is called a generalized pinball loss function-based TSVM (GPin-TSVM);
  • We demonstrate that the proposed algorithm GPin-TSVM surpasses existing classifiers in terms of accuracy in numerical experiments. We also examine its characteristics, such as noise sensitivity and within-class scatter;
  • We examine the applicability of the main techniques of GPin-TSVM toward handwritten digit recognition problems compared with the standard TSVM, Pin-TSVM, and ϵ -insensitive zone TSVM (IPin-TSVM). Moreover, we use the automatic feature extractor by the convolutional neural network (CNN) and TSVM, Pin-TSVM, IPin-SVM, and GPin-TSVM, which work as a binary classifier by replacing the softmax layer of CNN;
  • We perform numerical testing on a synthetic dataset and datasets from numerous UCI benchmarks with noise of various variances to illustrate the validity of our proposed GPin-TSVM. The results also show the robustness of the proposed approach, which is less sensitive to noise and retains the sparsity of the solution.
In Section 2, we briefly discuss loss functions, SVM, generalized pinball SVM, and TSVM. In Section 3, we present a new approach called GPin-TSVM. In Section 4, the properties of the proposed GPin-TSVM are discussed. The efficiency of our proposed GPin-TSVM by using synthetic datasets and the UCI machine learning repository is compared to standard TSVM, Pin-TSVM, and IPin-TSVM, and the applications of the proposed GPin-TSVM algorithms in handwritten digit recognition are shown in Section 5. Conclusions and future recommendations are presented in Section 6.

2. Related Work and Background

In this section, standard SVM, TSVM, loss functions, and generalized pinball SVM formulations are briefly described. The interested readers are referred to [28,39,40,42] for a more detailed description.

2.1. Support Vector Machine

The difficulty of the SVM model lies in determining the optimal separating hyperplane, or maximal margin hyperplane, that best separates the two classes in order to generalize new data to obtain accurate classification predictions. Consider a two-class dataset of m data samples, where S = { ( x 1 , y 1 ) , ( x 2 , y 2 ) , , ( x m , y m ) } for i = 1 , 2 , , m , x i R n is the sample with the label y i { 1 , 1 } . SVM models receive a separating produced decision function w x + b = 0 , where w R n and b R from the following problem:
min w , b 1 2 w 2 + C i = 1 m ξ i s . t . 1 y i ( w x i + b ) 1 ξ i , ξ i 0 , i = 1 , , m ,
where ξ i are the slack variables and C is the trade-off parameter. We obtain its dual QPP as follows by using the Lagrangian multipliers α i :
min α 1 2 i = 1 m j = 1 m y i y j ( x i x j ) α i α j i = 1 m α i s . t . i = 1 m y i α i = 0 , 0 α i C , i = 1 , , m .
We use the number of support vectors that satisfy 0 < α < C , represented by N S V , and we obtain the following decision function after optimizing this dual QPP:
y = sign i = 1 N S V α i * y i ( x i x ) + b
where α * denotes the dual problem solution (2).

2.2. Twin Support Vector Machine

Consider the data set S, in which the matrix A R m 1 × n represents m 1 data samples from class + 1 , and the matrix B R m 2 × n represents m 2 data samples from class 1 . The TSVM [28] is used to determine two nonparallel hyperplanes using the following definitions:
x w ( 1 ) + b ( 1 ) = 0 and x w ( 2 ) + b ( 2 ) = 0
where w ( 1 ) , w ( 2 ) R n and b ( 1 ) , b ( 2 ) R . To obtain the pair of nonparallel hyperplanes, the hinge loss function-based TSVM yields the following pair of QPPs:
min w ( 1 ) , b ( 1 ) , ξ 1 2 A w ( 1 ) + e 1 b ( 1 ) 2 + c 1 e 2 ξ s . t . ( B w ( 1 ) + e 2 b ( 1 ) ) + ξ e 2 , ξ 0 , and
min w ( 2 ) , b ( 2 ) , ξ 1 2 B w ( 2 ) + e 2 b ( 2 ) 2 + c 2 e 1 ξ s . t . ( A w ( 2 ) + e 1 b ( 2 ) ) + ξ e 1 , ξ 0 ,
where c 1 and c 2 are positive penalty parameters, ξ is a slack variable, and e 1 and e 2 are vectors of appropriately sized ones. The dual of QPPs (5) and (6) can be represented, respectively, as follows:
min α 1 2 α Q ( P P ) Q α e 2 α s . t . 0 α c 1 e 2
and
min β 1 2 β P ( Q Q ) P β e 1 β s . t . 0 β c 2 e 1
where P = A e 1 and Q = B e 2 . From the solutions α and β of (7) and (8), respectively, the best separating hyperplanes are given by:
w ( 1 ) b ( 1 ) = ( P P ) 1 Q α ,
and
w ( 2 ) b ( 2 ) = ( Q Q ) 1 P β .
Depending on which of the two hyperplanes (4) a new sample point x R n lies closest to, it is is assigned to class i ( i = + 1 or 1 ) by
class ( i ) = arg min i = 1 , 2 | x w ( i ) + b ( i ) | w ( i )
where | . | denotes obtaining the absolute value. In actuality, the unconstrained optimization problem may be reformulated as the QPP of the TSVM problem (5) as follows [43]:
min w ( 1 ) , b ( 1 ) , ξ 1 2 A w ( 1 ) + e 1 b ( 1 ) 2 + c 1 i = 1 m 2 L h i n g e 1 + ( x i w ( 1 ) + b ( 1 ) )
and
min w ( 2 ) , b ( 2 ) , ξ 1 2 B w ( 2 ) + e 2 b ( 2 ) 2 + c 2 i = 1 m 1 L h i n g e 1 ( x i w ( 2 ) + b ( 2 ) ) ,
where L h i n g e ( u ) = max 0 , u is known as the hinge loss function and u = ( 1 y i ( x i w + b ) ) . The hinge loss is a loss function that is commonly used to train classifiers. Furthermore, it strives to optimize the shortest distance between two classes, resulting in resampling instability and noise sensitivity from the related classifier [42]. To deal with the problem of noise sensitivity, Huang [39] presented utilizing the pinball loss function by combining the SVM classifier with the pinball loss function. The pinball loss function explains how this approach works by penalizing correctly identified data as follows:
L p i n ( u ) = u , u 0 , τ u , u < 0 ,
where τ 0 is a user-defined parameter. The so-called pinball loss is a well-known method in statistics and machine learning for calculating conditional quantiles. Despite achieving noise insensitivity, the pinball loss function is unable to achieve sparsity in the process. In their work, Huang examined a similar type of pinball loss function to ensure sparsity, which is a ϵ -insensitive pinball loss. The use of the ϵ -insensitive pinball loss function increases the prediction performance of the SVM model significantly. It also maintains sparsity in the SVM model. This function is defined as follows:
L p i n ϵ ( u ) = u ϵ , u > ϵ , 0 , ϵ τ u ϵ , τ u + ϵ τ , u < ϵ τ ,
where τ 0 and ϵ 0 are user-defined parameters. In an SVM model, sparsity is well known to be a highly desirable property. A sparse SVM model constructs the decision function from a small number of training data points and predicts the responses of test data points in a very short amount of time. The width of the ϵ -insensitive zone function fluctuates with the τ values, but it should ideally change with the variation in the training data response values. In practicality, it also makes choosing a good ϵ -value difficult. As a result, Rastogi [40] saw the necessity to create an ϵ -insensitive pinball loss function that can be used to enhance the ϵ -insensitive method in SVM. They proposed an ( ϵ 1 , ϵ 2 ) -insensitive zone pinball loss function by used this loss in combination with the SVM model. It is also called a generalized pinball SVM, with the following loss function:
L τ 1 , τ 2 ϵ 1 , ϵ 2 ( u ) = τ 1 u ϵ 1 τ 1 , u > ϵ 1 τ 1 , 0 , ϵ 2 τ 2 u ϵ 1 τ 1 , τ 2 u + ϵ 2 τ 2 , u < ϵ 2 τ 2 ,
where τ 1 , τ 2 , ϵ 1 , and ϵ 2 are non-negative parameters. In the next subsection, we briefly describe the generalized pinball SVM model that is proposed by Rastogi. This approach is a modification of previous loss functions that takes noise sensitivity, resampling stability, and data scatter minimization into account.

2.3. Support Vector Machine with Generalized Pinball Loss

With this generalized pinball loss function, the resulting formulation, termed as a generalized pinball support vector machine, is proposed by Rastogi [40], which results in the unconstrained optimization problem:
min w , b 1 2 w 2 + C m i = 1 m L τ 1 , τ 2 ϵ 1 , ϵ 2 1 y i ( w x i + b ) .
Then, the problem (15) can reformulate to the following QPP:
min w , b , ξ 1 2 w 2 + C i = 1 m ξ i s . t . y i ( w T x i + b ) 1 1 τ 1 ( ξ i + ϵ 1 ) , y i ( w T x i + b ) 1 + 1 τ 2 ( ξ i + ϵ 2 ) , ξ i 0 , i = 1 ,   2 ,   3 , , m .
Its dual QPP is generated as follows by inserting the Lagrangian multipliers α i and β i :
min α , β 1 2 i = 1 m j = 1 m ( α i β i ) ( α j β j ) y i y j ( x i x j ) ( 1 ϵ 1 τ 1 ) i = 1 m α i + ( 1 + ϵ 2 τ 2 ) i = 1 m β i s . t . i = 1 m ( α i β i ) y i = 0 , 0 α i τ 1 + β i τ 2 C , i = 1 , , m .
We can obtain the decision function (18) by solving the dual problem of (17):
y = sign i = 1 m ( α i * β i * ) y i ( x i x ) + b .
However, for large-scale applications, the generalized pinball SVM has a high computing complexity and is quite slow. In the next section, we go after the heart of our proposed technique, which is to reduce the high computational complexity by proposing a TSVM with a generalized pinball loss function that is aimed toward the binary classification problem, and to present both the linear and nonlinear cases, as shown in Figure 1.

3. Proposed Twin Support Vector Machine with Generalized Pinball Loss (GPin-TSVM)

In this section, we employ the Lagrange multiplier approach to derive the solution for our GPin-TSVM model, which is based on just the generalized pinball loss function. In both linear and nonlinear scenarios, our GPin-TSVM can be employed.

3.1. Linear Case

In the standard TSVM, we determine the generalized pinball loss and obtain the following QPPs:
min w ( 1 ) , b ( 1 ) 1 2 A w ( 1 ) + e 1 b ( 1 ) 2 + c 1 e 2 L τ 1 , τ 2 ϵ 1 , ϵ 2 e 2 + ( B w ( 1 ) + e 2 b ( 1 ) )
and
min w ( 2 ) , b ( 2 ) 1 2 B w ( 2 ) + e 2 b ( 2 ) 2 + c 2 e 1 L τ 3 , τ 4 ϵ 3 , ϵ 4 e 1 ( A w ( 2 ) + e 1 b ( 2 ) ) .
The problems (19) and (20) are translated further into equivalent known formulations (5) and (6) by adding a slack vector ξ , yielding the following QPPs:
min w ( 1 ) , b ( 1 ) , ξ 1 2 A w ( 1 ) + e 1 b ( 1 ) 2 + c 1 e 2 ξ s . t . ( B w ( 1 ) + e 2 b ( 1 ) ) e 2 1 τ 1 ( ξ + e 2 ϵ 1 ) , ( B w ( 1 ) + e 2 b ( 1 ) ) e 2 + 1 τ 2 ( ξ + e 2 ϵ 2 ) , ξ 0 ,
and
min w ( 2 ) , b ( 2 ) , ξ 1 2 B w ( 2 ) + e 2 b ( 2 ) 2 + c 2 e 1 ξ s . t . A w ( 2 ) + e 1 b ( 2 ) e 1 1 τ 3 ( ξ + e 1 ϵ 3 ) , A w ( 2 ) + e 1 b ( 2 ) e 1 + 1 τ 4 ( ξ + e 1 ϵ 4 ) , ξ 0 ,
where τ 1 , τ 2 , τ 3 , τ 4 , ϵ 1 , ϵ 2 , ϵ 3 , and ϵ 4 are non-negative parameters. We transform (21) and (22) to their dual form to arrive at the solution. For this, we use (21) and introduce the Lagrange multipliers α 0 , β 0 , and γ 0 , and we obtain the Lagrange function:
L ( w ( 1 ) , b ( 1 ) , ξ , α , β , γ ) = 1 2 A w ( 1 ) + e 1 b ( 1 ) 2 + c 1 e 2 ξ α ( B w ( 1 ) + e 2 b ( 1 ) ) e 2 + 1 τ 1 ( ξ + e 1 ϵ 1 ) β ξ γ ( B w ( 1 ) + e 2 b ( 1 ) ) + e 2 + 1 τ 2 ( ξ + e 2 ϵ 2 ) .
We use the KKT optimality conditions to find the following results:
L w ( 1 ) = A ( A w ( 1 ) + e 1 b ( 1 ) ) + B α B γ = 0 ,
L b ( 1 ) = e 1 ( A w ( 1 ) + e 1 b ( 1 ) ) + e 2 α e 2 γ = 0 ,
L ξ = c 1 e 2 α τ 1 β γ τ 2 = 0 ,
α ( B w ( 1 ) + e 2 b ( 1 ) ) e 2 + 1 τ 1 ( ξ + e 1 ϵ 1 ) = 0 ,
β ξ = 0 ,
γ ( B w ( 1 ) + e 2 b ( 1 ) ) + e 2 + 1 τ 2 ( ξ + e 2 ϵ 2 ) = 0 .
By using (26) and β 0 , we obtain
α τ 1 + γ τ 2 c 1 e 2
Combining (24) and (25) yields
A e 1 A e 1 w ( 1 ) b ( 1 ) B e 2 ( α γ ) = 0 .
Define λ = α γ , P = A e 1 , and Q = B e 2 . Equation (31) can be recast using these notations as follows:
P P w ( 1 ) b ( 1 ) + Q λ = 0 , i . e . , w ( 1 ) b ( 1 ) = ( P P ) 1 Q λ .
We can obtain the dual of (21) using Equation (23) and the given KKT conditions as follows:
min α , λ 1 2 λ Q ( P P ) 1 Q λ λ e 2 ( 1 + ϵ 2 τ 2 ) + α e 2 ϵ 1 τ 1 + ϵ 2 τ 2 s . t . 0 1 τ 1 + 1 τ 2 α λ τ 2 c 1 e 2 , α 0 , α λ 0 .
The dual problem of (22) can be derived similarly:
min ω , μ 1 2 μ P ( Q Q ) 1 P μ μ e 1 ( 1 + ϵ 4 τ 4 ) + ω e 1 ϵ 3 τ 3 + ϵ 4 τ 4 s . t . 0 1 τ 3 + 1 τ 4 ω μ τ 4 c 2 e 1 , ω 0 , μ 0 ,
where ω 0 and μ 0 are Lagrange multipliers. Finally, the best separating hyperplanes are given by:
w ( 1 ) b ( 1 ) = ( P P + δ I ) 1 Q λ ,
and
w ( 2 ) b ( 2 ) = ( Q Q + δ I ) 1 P μ .
Since we cannot ensure P P and Q Q are irreversible, it is always positive semi-definite; however, in some circumstances, it may not be well conditioned. To account for the possibility of ill-conditioning of P P and Q Q , the regularization term δ I ( δ > 0 ) must be used [44]. Depending on which of the two hyperplanes (4) a new sample point x R n lies closest to, it is assigned to class i ( i = + 1 or −1) by
class ( i ) = arg min i = 1 , 2 | x w ( i ) + b ( i ) | w ( i ) .

3.2. Nonlinear Case

In higher dimensions, support vector machines are even more difficult to interpret. It is considerably more difficult to view how the data can be separated linearly and what the decision boundary will look like. In practice, however, data are rarely linearly separable; therefore, we must transform it into a higher-dimensional space before developing a support vector classifier. This problem can be solved using the symmetric kernel trick. Now, we use a symmetric kernel method to extend our linear GPin-TSVM to the nonlinear case [28,45]. The symmetric kernels used have a significant impact on how well GPin-TSVM functions. The nonparallel hyperplanes in the kernel-generated space are as follows if the defined kernel function is K ( · , · ) :
K ( x , X ) w ( 1 ) + b ( 1 ) = 0 and K ( x , X ) w ( 2 ) + b ( 2 ) = 0 ,
where w ( 1 ) , w ( 2 ) R m , and X = A m 1 × n B m 1 × n . For the nonlinear case of the problems (21) and (22), the corresponding problems are
min w ( 1 ) , b ( 1 ) , ξ 1 2 K ( A , X ) w ( 1 ) + e 1 b ( 1 ) + c 1 e 2 ξ s . t . ( K ( B , X ) w ( 1 ) + e 2 b ( 1 ) ) e 2 1 τ 1 ( ξ + e 2 ϵ 1 ) , ( K ( B , X ) w ( 1 ) + e 2 b ( 1 ) ) e 2 + 1 τ 2 ( ξ + e 2 ϵ 2 ) , ξ 0 ,
and
min w ( 2 ) , b ( 2 ) , ξ 1 2 K ( B , X ) w ( 2 ) + e 2 b ( 2 ) + c 2 e 1 ξ s . t . K ( A , X ) w ( 2 ) + e 1 b ( 2 ) e 1 1 τ 3 ( ξ + e 1 ϵ 3 ) , K ( A , X ) w ( 2 ) + e 1 b ( 2 ) e 1 + 1 τ 4 ( ξ + e 1 ϵ 4 ) , ξ 0 .
The Lagrange function is applied, and the KKT optimality requirements are used to produce the dual of (37):
min α , γ 1 2 ( α γ ) Q ( P P ) 1 Q ( α γ ) ( α γ ) e 2 ( 1 + ϵ 2 τ 2 ) + α e 2 ϵ 1 τ 1 + ϵ 2 τ 2 s . t . α τ 1 + γ τ 2 c 1 e 1 , α 0 , α γ 0 .
Similarly, the dual of Equation (38) can be obtained as follows:
min ω , μ 1 2 ( ω μ ) P ( Q Q ) 1 P ( ω μ ) ( ω μ ) e 1 ( 1 + ϵ 4 τ 4 ) + ω e 1 ϵ 3 τ 3 + ϵ 4 τ 4 s . t . ω τ 3 + μ τ 4 c 2 e 1 , ω 0 , ω μ 0 ,
where P = K ( A , X ) e 1 , Q = K ( B , X ) e 2 , and α , γ , ω , and μ are Lagrange multipliers. Finally, the best separating hyperplanes are given by:
w ( 1 ) b ( 1 ) = ( P P + δ I ) 1 Q ( α γ )
and
w ( 2 ) b ( 2 ) = ( Q Q + δ I ) 1 P ( ω μ ) .
Thus, a new sample point x R n is assigned to class i ( i = + 1 or 1 ) by
class ( i ) = arg min i = 1 , 2 | K ( x , X ) w ( i ) + b ( i ) | w ( i ) .

4. Properties of the GPin-TSVM

We examine the noise insensitivity and within-class scatter properties of the GPin-TSVM in this section.

4.1. Noise Insensitivity

The principal advantage of our proposed algorithm GPin-TSVM is that it is insensitive to noise and maintains the sparsity. In this subsection, we explain the advantage of giving a penalty on a correctly classified point and conserving the sparsity to a certain scale at the same time. Consider the generalized sign function sgn τ 1 , τ 2 ϵ 1 , ϵ 2 ( 1 y ( w x + b ) ) as
sgn τ 1 , τ 2 ϵ 1 , ϵ 2 ( u ) = { τ 1 } , u > ϵ 1 τ 1 , [ 0 , τ 1 ] , u = ϵ 1 τ 1 , { 0 } , ϵ 2 τ 2 < u < ϵ 1 τ 1 , [ τ 2 , 0 ] , u = ϵ 2 τ 2 , { τ 2 } , u < ϵ 2 τ 2 .
sgn τ 1 , τ 2 ϵ 1 , ϵ 2 ( u ) is the subgradient of (14). In the linear case, we will concentrate on the first model of the GPin-TSVM for clarity. Using the KKT optimality condition, Equation (19) can be written as:
0 A ( A w ( 1 ) + e 1 b ( 1 ) ) + c 1 i = 1 m 2 sgn τ 1 , τ 2 ϵ 1 , ϵ 2 ( 1 + ( w ( 1 ) x i + b ( 1 ) ) ) x i ,
where 0 is a zero vector. For the given w ( 1 ) and b ( 1 ) , the entire index set can be divided into five different subsets:
E 1 + = i : 1 + ( w ( 1 ) x i + b ( 1 ) ) > ϵ 1 τ 1 , E 2 + = i : 1 + ( w ( 1 ) x i + b ( 1 ) ) = ϵ 1 τ 1 , E 3 + = i : ϵ 2 τ 2 < 1 + ( w ( 1 ) x i + b ( 1 ) ) < ϵ 1 τ 1 , E 4 + = i : 1 + ( w ( 1 ) x i + b ( 1 ) ) = ϵ 2 τ 2 , E 5 + = i : 1 + ( w ( 1 ) x i + b ( 1 ) ) < ϵ 2 τ 2 .
The data samples in E 3 + may not benefit w ( 1 ) because the sub-gradient at all these datasets is zero, which is shown in Equation (42). As a result, E 3 + has a direct impact on the model sparsity. We perceive that ϵ 1 and ϵ 2 control the number of samples in E 3 + . As ϵ 1 and ϵ 2 approach 0, sparsity is lost, whereas if ϵ 1 and ϵ 2 , we increase the sparsity as a consequence of having more samples in E 3 + .
Using the notation E 1 + , E 2 + , E 3 + , E 4 + , and E 5 + , Equation (43) can be rewritten as the existence of ψ i [ 0 , τ 1 ] and θ i [ τ 2 , 0 ] , such that
1 c 1 A ( A w ( 1 ) + e 1 b ( 1 ) ) + τ 1 i E 1 + x i + i E 2 + ψ i x i + i E 4 + θ i x i τ 2 i E 5 + x i = 0
where i = 1 , , m 2 .
Theorem 1.
Let p 1 be the number of samples x i in E 1 + . The following inequalities must hold if the optimization problems (33) or (39) have a solution:
τ 1 + e 1 ( A w ( 1 ) + e 1 b ( 1 ) ) c 1 m 2 0
and
p 1 m 2 1 τ 1 + e 1 ( A w ( 1 ) + e 1 b ( 1 ) ) c 1 m 2 τ 1 + τ 2 .
Proof. 
Let x i 0 be an arbitrary sample in E 1 + . We have β i 0 = γ i 0 = 0 by using the KKT condition (28) and (29). We obtain α i 0 = c 1 τ 1 by using the KKT condition (26), which implies that α i 0 γ i 0 = c 1 τ 1 . Let λ = α γ , which implies that λ i 0 = c 1 τ 1 . In addition, we obtain e 1 ( A w ( 1 ) + e 1 b ( 1 ) ) = p 1 γ i 0 + i E 1 + λ i = p 1 c 1 τ 1 + i E 1 + λ i from the KKT condition (25).
We can obtain c 1 τ 2 λ i c 1 τ 1 because α i 0 and γ i 0 . As a result, we obtain
e 1 ( A w ( 1 ) + e 1 b ( 1 ) ) τ 1 c 1 ( m 2 p 1 ) p 1 c 1 τ 1 e 1 ( A w ( 1 ) + e 1 b ( 1 ) ) + τ 2 c 1 ( m 2 p 1 ) ,
thus, e 1 ( A w ( 1 ) + e 1 b ( 1 ) ) c 1 τ 1 m 2 1 and p 1 1 + τ 2 τ 1 e 1 ( A w ( 1 ) + e 1 b ( 1 ) ) + τ 2 c 1 m 2 c 1 τ 1 . Finally, we have
p 1 m 2 1 c 1 ( τ 1 + τ 2 ) e 1 ( A w ( 1 ) + e 1 b ( 1 ) ) m 2 + τ 2 c 1 = 1 τ 1 + e 1 ( A w ( 1 ) + e 1 b ( 1 ) ) c 1 m 2 τ 1 + τ 2 .
Theorem 1 implies that 1 τ 1 + e 1 ( A w ( 1 ) + e 1 b ( 1 ) ) c 1 m 2 τ 1 + τ 2 is an upper boundary of the number of samples in E 1 + . The parameters τ 1 and τ 2 control the numbers of samples in E 1 + , E 3 + , and E 5 + . When there is a decrease in τ 1 and τ 2 , then the number of elements in E 1 + becomes smaller and the classification result is sensitive to feature noise around the decision boundary, which will have a considerable impact. When τ 1 and τ 2 are both large, all three sets contain a large number of samples, making the outcome less sensitive to feature noise.
Briefly, parameters ϵ 1 , ϵ 2 , τ 1 , and τ 2 control the tradeoff between sparsity and noise insensitivity. Similarly, we can separate the index set into the five sets on the second model of the GPin-TSVM:
E 1 = i : 1 ( w ( 2 ) x i + + b ( 2 ) ) > ϵ 3 τ 3 , E 2 = i : 1 ( w ( 2 ) x i + + b ( 2 ) ) = ϵ 3 τ 3 , E 3 = i : ϵ 4 τ 4 < 1 ( w ( 2 ) x i + + b ( 2 ) ) < ϵ 3 τ 3 , E 4 = i : 1 ( w ( 2 ) x i + + b ( 2 ) ) = ϵ 4 τ 4 , E 5 = i : 1 ( w ( 2 ) x i + + b ( 2 ) ) < ϵ 4 τ 4
where i = 1 , , m 1 . Similar properties of the parameters τ 3 and τ 4 can be obtained as follows:
Theorem 2.
Let p 2 be the number of samples x i + in E 1 . If the optimization problems (34) or (40) have a solution, then the following inequalities must hold:
τ 3 e 2 ( B w ( 2 ) + e 2 b ( 2 ) ) c 2 m 1 0
and
p 2 m 1 1 τ 3 e 2 ( B w ( 2 ) + e 2 b ( 2 ) ) c 2 m 1 τ 3 + τ 4 .
It also indicates that 1 τ 3 e 2 ( B w ( 2 ) + e 2 b ( 2 ) ) c 2 m 1 τ 3 + τ 4 is an upper bound on the number of samples in E 1 .

4.2. Scatter Minimization

Scatter minimization can also be used to understand the GPin-TSVM. For simplicity, consider only the first QPP (19) of the GPin-TSVM. The conclusions for another QPP (20) can also be obtained in this manner. For the given x i B and x j + A , the positive hyperplane x w ( 1 ) + b ( 1 ) = 0 can be established by data samples under the subset Y 2 + A , and the two hyperplanes H + = { w ( 1 ) x i + b ( 1 ) + 1 = 0 } and H = { w ( 2 ) x j + + b ( 2 ) 1 = 0 } are defined by data samples in subsets Y 3 + E 3 + and subset Y 3 E 3 , respectively.
The scatter is calculated by adding the distances between each point x i and one supplied negative sample x i 3 Y 3 + . The scatter of x i B around the sample x i 3 can be determined as
i = 1 m 2 | w ( 1 ) x i 3 + b ( 1 ) ( w ( 1 ) x i + b ( 1 ) ) | = i = 1 m 2 | w ( 1 ) ( x i 3 x i ) | .
We obtain the following equation by using w ( 1 ) x i 3 + b ( 1 ) + 1 = 0 :
i = 1 m 2 | w ( 1 ) ( x i 3 x i ) | = i = 1 m 2 | w ( 1 ) x i 3 + b ( 1 ) ( w ( 1 ) x i + b ( 1 ) ) | = i = 1 m 2 | 1 ( w ( 1 ) x i + b ( 1 ) ) | = i = 1 m 2 | 1 + ( w ( 1 ) x i + b ( 1 ) ) | .
Similarly, using a specific data sample x j 2 + Y 2 + , the scatter for every sample x j + A is calculated as follows:
j = 1 m 1 | w ( 1 ) ( x j 2 + x j + ) | = j = 1 m 1 | w ( 1 ) x j 2 + + b ( 1 ) ( w ( 1 ) x j + + b ( 1 ) ) | = i = 1 m 1 | ( w ( 1 ) x j + + b ( 1 ) ) |
where w ( 1 ) x j 2 + + b ( 1 ) = 0 . Due to the fact that the scatter is a positive value, we can take it as the sum of squares, i.e., i = 1 m 1 ( ( w ( 1 ) x j + + b ( 1 ) ) ) 2 .
Consider the formula as follows:
min w ( 1 ) , b ( 1 ) 1 2 i = 1 m 1 ( ( w ( 1 ) x j + + b ( 1 ) ) ) 2 + c 11 i = 1 m 2 | 1 + ( w ( 1 ) x i + b ( 1 ) ) |
where c 11 is a constant. This guarantees that the first term may be expressed in such a way that the scatters x j + A about the hyperplane x w ( 1 ) + b ( 1 ) = 0 are minimized. Nevertheless, this second term seeks to lower the error values caused according to how close B samples must be to H + by minimizing the scatter of x i B from around hyperplane H + .
In the GPin-TSVM (19), the first term of (45) is mentioned in its mathematically equivalent form, whereas the absolute value used in (45) is extended to L τ 1 , τ 2 ϵ 1 , ϵ 2 .
The first term of (45) is expressed within GPin-TSVM (19) in its mathematically equivalent form, whereas the absolute value employed in (45) is extended to L τ 1 , τ 2 ϵ 1 , ϵ 2 . Concretely, we introduce the misclassification term
c 12 L h i n g e 1 + ( w ( 1 ) x i + b ( 1 ) ) ϵ 1 τ 1 = c 12 max 0 , 1 + ( w ( 1 ) x i + b ( 1 ) ) ϵ 1 τ 1 c 13 L h i n g e ( 1 + ( w ( 1 ) x i + b ( 1 ) ) ) = c 13 max { 0 , 1 + ( w ( 1 ) x i + b ( 1 ) ) } c 14 L h i n g e ( 1 ( w ( 1 ) x i + b ( 1 ) ) ) = c 14 max { 0 , 1 ( w ( 1 ) x i + b ( 1 ) ) } c 15 L h i n g e 1 ( w ( 1 ) x i + b ( 1 ) ) ϵ 2 τ 2 = c 15 max 0 , 1 ( w ( 1 ) x i + b ( 1 ) ) ϵ 2 τ 2
into (45), where c 12 , c 13 , c 14 and c 15 are positive parameters; that is,
min w ( 1 ) , b ( 1 ) 1 2 i = 1 m 1 ( ( w ( 1 ) x j + + b ( 1 ) ) ) 2 + c 11 i = 1 m 2 | 1 + ( w ( 1 ) x i + b ( 1 ) ) | + c 12 i = 1 m 2 L h i n g e ( 1 + ( w ( 1 ) x i + b ( 1 ) ) ϵ 1 τ 1 ) + c 13 i = 1 m 2 L h i n g e ( 1 + ( w ( 1 ) x i + b ( 1 ) ) ) + c 14 i = 1 m 2 L h i n g e ( 1 ( w ( 1 ) x i + b ( 1 ) ) ) + c 15 i = 1 m 2 L h i n g e 1 ( w ( 1 ) x i + b ( 1 ) ) ϵ 2 τ 2 .
The GPin-TSVM (19) can be obtained using the following conditions: c 11 + c 12 + c 13 = c 1 τ 1 , c 11 + c 13 = 0 , c 11 + c 14 = 0 , and c 11 + c 14 + c 15 = c 1 τ 2 . We have τ 1 = c 12 c 1 and τ 2 = c 15 c 1 from the first and last condition. The interpretation of this report is that the reasonable range of τ 1 0 and τ 2 0 . We examine the misclassification error and the within-class scatter of one class simultaneously in the generalized pinball loss minimization. The GPin-TSVM (19) is then regarded as a trade-off between low misclassification and reduced scatter.

5. Numerical Experiments

In this section, the classification performance of the proposed approach in terms of accuracy is compared to that of other relevant approaches, such as the hinge loss twin support vector machine (TSVM), pinball loss TSVM (Pin-TSVM), and ϵ - insensitive loss TSVM (IPin-TSVM), on synthetic datasets and the UCI machine learning repository [46], and handwritten digit recognition applications have been proposed. We employed 10-fold cross validation for all of our experiments. The average accuracy and standard deviation for each experiment are displayed in all tables, with the best one highlighted.
All experiments are implemented in Python 3.9.5. on Windows 8 running on a 1.9 GHz laptop with 4 GB RAM with system configuration Intel Core i5+ Duo CPU E7500 (2.93 GHz). From now on, we denote τ 1 = τ 3 , τ 2 = τ 4 , ϵ 1 = ϵ 3 , and ϵ 2 = ϵ 4 . To derive the nonlinear case, we use the radial basis function kernel K ( x , y ) = exp { x y 2 2 σ } .

5.1. Synthetic Dataset

We test our approach on a two-dimensional case in which equal samples are drawn from two Gaussian distributions: x i , i { i : y i = 1 } N ( μ 1 , Σ 1 ) and x i , i { i : y i = 1 } N ( μ 2 , Σ 2 ) , where μ 1 = [ 1 , 3 ] , μ 2 = [ 1 , 3 ] and Σ 1 = Σ 2 = 0.2 0 0 3 . In order to make the dataset more interesting, we introduce noise. The labels of the noise points are chosen with equal probabilities from { 1 , 1 } . The placements of these samples match the Gaussian distribution N ( μ n , Σ n ) , where μ n = [ 0 , 0 ] and Σ n = 1 0.8 0.8 1 . The labels around the decision boundaries are affected by the noise. The ratio of noise data in the training set is represented by r. From Figure 2, the bar chart demonstrates the percentages of the accuracy of classifying different sectors, including GPin-TSVM, IPin-TSVM, Pin-TSVM, and TSVM during r = 0 % to 30 % . In the large majority of cases, the GPin-TSVM produces the greatest outcome. This implies that the GPin-TSVM was the strongest candidate for the method of classifying the noise-corrupted data.
In the next result in Figure 3, we show the obtained value of the slopes of hyperplanes over four different noisy synthetic datasets by SVM, TSVM, and the proposed GPin-TSVM. In this result, we show that, when the level of noise increases from 0 to 20%, the hyperplanes of SVM diverge from 0.6575 to 0.1592 and the hyperplanes of TSVM diverge from 1.7988, 1.2688 to 0.2879, 0.2562, whereas the hyperplanes of our GPin-TSVM slightly changes. This suggests that our proposed GPin-TSVM model is unaffected by noise near the boundary.

5.2. UCI Datasets

Additionally, we perform testing on 10 benchmark datasets from the UCI machine learning database [46]. Imbalanced datasets lead to incorrect classification in classification problems. The imbalance ratio (IR) [47] is defined as the ratio of the number of data points on the majority class to the number of data points on the minority class.
IR = number of data points on the majority class number of data points on the minority class .
The dataset descriptions can be found in Table 1. To modify the tradeoff parameters and kernel parameter σ for UCI benchmark datasets, we used the grid search method [48]. A validation set of 10% randomly selected data points was used for each dataset. We chose values for parameters c 1 and c 2 from the set { 10 i | i = 2 , 1 , 0 , 1 , 2 } for our tests. Further, another parameter was tuned in the range { 0.1 , 0.25 , 0.5 , 0.75 , 1 , 1.5 , 2 , 2.5 } . Table 2 and Table 3 summarize the experimental results of four approaches (TSVM, Pin-TSVM, IPin-SVM, and GPin-TSVM) on linear and RBF kernels, respectively. The optimal parameters used in Table 2 and Table 3 are summarized in Table 4 and Table 5, respectively. Accuracy is defined as the mean value of ten time-testing results plus or minus the standard deviation in Table 2 and Table 3.
Table 2 illustrates the results of applying TSVM, Pin-TSVM, IPin-TSVM, and our proposed GPin-TSVM to a linear kernel on eight distinct UCI datasets. The results with the highest accuracy are highlighted in bold. In most datasets, the classification performance of GPin-TSVM outperforms TSVM, Pin-TSVM, and IPin-TSVM in terms of accuracy, according to the experimental results. Our proposed GPin-TSVM has the highest prediction accuracy in 20 of the 32 scenarios. Furthermore, when the number of noise samples varies from r = 0 (noise free) to r = 0.2 , our proposed GPin-TSVM outperforms existing methods in terms of classification accuracy and stability. However, in Breast, the classification accuracies of IPin-TSVM are better than those of our proposed GPin-TSVM.
The nonlinear kernel with an RBF kernel was subjected to a similar analysis, with the results presented in Table 3. Our proposed GPin-TSVM has the best prediction accuracy in 20 of the 32 cases. In most of the datasets, our proposed GPin-TSVM offers the best prediction accuracy, as shown in Table 2 and Table 3. As a result, the accuracy of our proposed GPin-TSVM outperforms that of existing models.
The sparsity of the proposed approach of the GPin-TSVM is compared to that of the standard TSVM for the linear and nonlinear cases in Table 6 and Table 7, respectively. When we look at the results, we can see that as ϵ 1 = ϵ 2 grows, our solution becomes more sparse. It is clear from both tables that our proposed GPin-TSVM is more sparse than the standard TSVM while still keeping noise-insensitive properties. The prediction process is faster than the standard TSVM because of the sparsity of the solution, which is extremely useful in datasets with big samples.

5.3. Hybrid CNN-GPin-TSVM Classifier for Handwritten Digit Recognition

The proposed algorithms GPin-TSVM and their application to handwritten digit recognition problems are discussed in this part. Handwritten digit recognition is a difficult topic that has been intensively researched in the subject of handwriting recognition for many years. As a result of its many practical uses and financial implications, handwritten digit recognition is still a popular topic. Here, we use MNIST handwritten datasets to carry out the experiments. In the field of machine learning, the MNIST dataset is commonly used for training and testing. There are 60,000 samples in the training set and 10,000 in the test set in this dataset. Each sample has a size of 28 × 28 pixels. As seen in Figure 4, the MNIST dataset comprises grayscale images of handwritten digits from ‘0’ to ‘9’. Several approaches for handwriting recognition have been proposed in the literature, such as k-nearest neighbor (KNN) [49], SVM [49,50,51,52], artificial neural network (ANN) [53,54], convolutional neural network (CNN) [51,55,56], etc.
One of the most important aspects of our cognition system success is feature extraction. Traditional feature extraction by hand is a tedious and time-consuming procedure that does not work with raw images, but features can be recovered directly from raw images using automatic extraction algorithms. On ear recognition, Alshazly [57] analyzed CNN-learnt features that are automatically optimized and found that features extracted by CNN produced the capacity to learn more specific features that are robust to wide image variations and to obtain a state-of-the-art recognition performance. On the clinical electroencephalogram (EEG) data classification problem, Xin [58] constructed a convolution support vector machine for classifying epilepsy EEG signals, and produced the highest accuracy. Recently, the hybrid CNN–SVM classifier for recognizing handwritten digits was proposed by [50,51,52]. They created a hybrid model that combines a powerful CNN with a SVM for handwritten digit recognition using the MNIST dataset, where SVM is a binary classifier and CNN is an automatic feature extractor, which both display a strong efficiency for handwritten digit recognition. Inspired by this particular work, the goal of this section is to use CNN to extract features from the MNIST dataset of input handwritten digit images. TSVM, Pin-TSVM, IPin-SVM, and GPin-TSVM work as a binary classifier, replacing the softmax layer of CNN. Moreover, we compare the performance between TSVM, Pin-TSVM, IPin-TSVM, and GPin-TSVM. We choose four pairs of handwritten digits on raw pixel features for our comparisons.
We build the network using the following. The first is placing the convolutional (Conv2D) layer into a channel of dimension 1, since the images are grayscale. The kernel size is set to 5 × 5 with a stride of 1. This convolution output is set to nine channels, implying that it will extract nine feature maps using nine kernels. We use a padding size of 1 to ensure that the input and output dimensions are the same. These layer output dimensions are 9 × 28 × 28 . The second convolutional (Conv2D) layer has an input channel size of 9. We set the output channel size to 16, which implies that 16 feature maps will be extracted. This layer kernel size is 5 with a stride of 1. After that, we add a RelU activation and a pooling (MaxPool2D) layer with a kernel of size 2 and a stride of 2. The pooling layer is mainly integrated to reduce the data dimension. Finally, two fully connected layers are used. The first fully connected layer will receive a flattened version of the feature maps. As a result, it must have a dimension of 16 × 7 × 7 , or 256 nodes. This layer will be connected to a fully connected 80-node layer. Finally, a hidden layer of the neural network containing 84 nodes is implemented. After the architecture of the model is defined, the model needs to be compiled. Here, we use TSVM, Pin-TSVM, IPin-SVM, and GPin-TSVM, as it is a binary classification problem. The architecture of the proposed model is described in Figure 5.
We compare the performance of the proposed model on MNIST handwritten datasets with other supervised recognition systems. The results of the suggested approach on the MNIST handwritten dataset are shown in Figure 6 and the optimal parameters of the result in Figure 6 are shown in Table 8. From Figure 6, we can learn that the classification performance of GPin-TSVM yields the best prediction accuracy of two pairwise digits out of the four total ones in terms of accuracy. On the 1 vs. 7 pairwise digit, our proposed GPin-TSVM and IPin-TSVM have an accuracy of 99.80%, which is greater than the recognition accuracy of another Pin-TSVM and TSVM classifier. However, the accuracy of GPin-TSVM on some pairwise digits, such as 5 vs. 8, is not the best. Overall, our GPin-TSVM performs well in terms of accuracy.

5.4. Statistical Analysis

On the four pairs of handwritten digits, the Friedman test is primarily used to evaluate the classification performance of the proposed GPin-TSVM algorithm. The Friedman test, along with post hoc testing, is a statistical test method that ranks algorithms differently for each data set, with the best method having the lowest ranking number [59]. The tests allow for a more accurate assessment of the algorithms’ relevance. We compare four different classifiers on four different pairs of handwritten digits. The accuracy of the related classifiers on each dataset is ranked, and the classifier with the highest accuracy has the smallest rank r i . Based on the accuracy of the four pairs of handwritten digits, the average rank of all methods is shown in Table 9.
Under the null hypothesis, the chi-square distribution and F-distribution with a degree of freedom ( k 1 ) ( N 1 ) in the Friedman test are:
X F 2 = 12 N k ( k + 1 ) j R j 2 k ( k + 1 ) 2 4 , F F = ( N 1 ) X F 2 N ( k 1 ) X F 2 ,
respectively, where R j = 1 N j = 1 N r j , the number of methods is k, and the number of datasets is N. According to Table 9, we obtain X F 2 = 10.58 and F F = 22.35 . For a significance level of 0.05 , the F ( 3 , 9 ) critical value is 3.86 , and 22.35 > 3.86 . As a result, the null hypothesis is rejected, i.e., there is a significant difference here between the four classifiers. Furthermore, as shown in Table 9, the proposed GPin-TSVM was ranked lowest on average. On the four pairs of handwritten digits, the classification performance of the proposed GPin-TSVM outperforms the other classifiers.

6. Conclusions

In this paper, a new version of a TSVM for pattern classification—a twin support vector machine—is created, and a generalized pinball loss function (GPin-TSVM) is implemented to improve the TSVM generalization performance, providing a lower sensitivity to noise and the ability to handle losing sparsity. We conduct wide experiments on synthetic datasets and the UCI machine learning repository, and handwritten digit recognition applications are compared to standard TSVM, Pin-TSVM, and IPin-TSVM. In most cases, the accuracy performance of our suggested GPin-TSVM is superior to that of existing classifiers, according to the experimental data. Additionally, the GPin-TSVM is less sensitive to noise and achieves sparsity, which is a major benefit of our proposed method. In the proposed GPin-TSVM, we also investigate the effect of value ϵ i ( i = 1 , 2 ) . GPin-TSVM is more sparse than standard TSVM for the sparsity of the solution. At last, the proposed algorithm of GPin-TSVM was used to solve the problem of handwritten digit recognition in the application, and we used the Friedman test to evaluate the classification performance of the proposed GPin-TSVM algorithm. From the results, it can be seen that our proposed GPin-TSVM is an effective approach for handwritten digit recognition, thereby demonstrating the effectiveness of the proposed algorithm.
Our future study will focus on the applicability of GPin-TSVM to multi-class supervised classification problems and to large-scale classification problems.

Author Contributions

Conceptualization, W.P. and R.W.; methodology, W.P. and R.W.; software, W.P. and W.R.; validation, W.P.; writing—original draft, W.P.; writing—review and editing, W.P. and R.W.; supervision, R.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the NSRF and NU, Thailand, with Grant Number R2564E044.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

The authors are thankful to the referees for their attentive reading and valuable suggestions. This research is partially supported by Development and Promotion of the Gifted in Science and Technology Project and Naresuan University.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Balasundaram, S.; Tanveer, M. On proximal bilateral-weighted fuzzy support vector machine classifiers. Int. J. Adv. Intell. Paradig. 2012, 4, 199–210. [Google Scholar] [CrossRef]
  2. Chang, F.; Guo, C.Y.; Lin, X.R.; Lu, C.J. Tree decomposition for large-scale SVM problems. J. Mach. Learn. Res. 2010, 11, 2935–2972. [Google Scholar]
  3. Zhang, C.; Tian, Y.; Deng, N. The new interpretation of support vector machines on statistical learning theory. Sci. China 2010, 53, 151–164. [Google Scholar] [CrossRef]
  4. Smola, A.J.; Scholkopf, B. A tutorial on support vector regression. Stat. Comput. 2004, 14, 199–222. [Google Scholar] [CrossRef] [Green Version]
  5. Tang, L.; Tian, Y.; Pardalos, P.M. A novel perspective on multiclass classification: Regular simplex support vector machine. Inf. Sci. 2019, 480, 324–338. [Google Scholar] [CrossRef]
  6. van de Wolfshaar, J.; Karaaba, M.F.; Wiering, M.A. Deep Convolutional Neural Networks and Support Vector Machines for Gender Recognition. In Proceedings of the IEEE Symposium Series on Computational Intelligence, Cape Town, South Africa, 7–10 December 2015; pp. 188–195. [Google Scholar]
  7. Lilleberg, J.; Zhu, Y.; Zhang, Y. Support vector machines and Word2vec for text classification with semantic features. In Proceedings of the IEEE 14th International Conference on Cognitive Informatics & Cognitive Computing (ICCI*CC), Beijing, China, 6–8 July 2015; pp. 136–140. [Google Scholar]
  8. Mohammad, A.H.; Alwada’n, T.; Al-Momani, O. Arabic Text Categorization Using Support vector machine, Naïve Bayes and Neural Network. GSTF J. Comput. 2016, 5, 108–115. [Google Scholar] [CrossRef]
  9. Mehmood, Z.; Mahmood, T.; Javid, M.A. Content-based image retrieval and semantic automatic image annotation based on the weighted average of triangular histograms using support vector machine. Appl. Intell. 2018, 48, 166–181. [Google Scholar] [CrossRef]
  10. Richhariya, B.; Tanveer, M. EEG signal classification using universum support vector machine. Expert Syst. Appl. 2018, 106, 169–182. [Google Scholar] [CrossRef]
  11. Soula, A.; Tbarki, K.; Ksantini, R.; Saida, S.B.; Lachiri, Z. A novel incremental Kernel Nonparametric SVM model (iKN-SVM) for data classification: An application to face detection. Eng. Appl. Artif. Intell. 2020, 89, 103468. [Google Scholar] [CrossRef]
  12. Krishna, G.; Prakash, N. A new training approach based on ECOC-SVM for SAR image retrieval. Int. J. Intell. Enterp. 2021, 8, 492–517. [Google Scholar] [CrossRef]
  13. Jayadeva; Khemchandani, R.; Chandra, S. Twin Support Vector Machines: Models; Springer: Cham, Switzerland, 2016. [Google Scholar]
  14. Xu, J.; Xu, C.; Zou, B.; Tang, Y.Y.; Peng, J.; You, X. New Incremental Learning Algorithm With Support Vector Machines. IEEE Trans. Syst. 2018, 49, 2230–2241. [Google Scholar] [CrossRef]
  15. Catak, F.Ö. Classification with boosting of extreme learning machine over arbitrarily partitioned data. Soft Comput. 2017, 21, 2269–2281. [Google Scholar] [CrossRef] [Green Version]
  16. Jayadeva; Khemchandani, R.; Chandra, S. Twin support vector machines for pattern classification. IEEE Trans. Pattern Anal. Mach. Intell. 2007, 29, 905–910. [Google Scholar] [CrossRef] [PubMed]
  17. Kumar, M.; Gopal, M. Least squares twin support vector machines for text categorization. In Proceedings of the 39th National Systems Conference (NSC), Greater Noida, India, 14–16 December 2015. [Google Scholar]
  18. Francis, L.M.; Sreenath, N. Robust Scene Text Recognition: Using Manifold Regularized Twin-SupportVector Machine. J. King Saud Univ. Comput. Inf. Sci. 2007. Available online: https://www.sciencedirect.com/science/article/pii/S1319157818309509 (accessed on 2 February 2019).
  19. Agarwal, S.; Tomar, D. Siddhant Prediction of software defects using Twin Support Vector Machine. In Proceedings of the International Conference on Information Systems and Computer Networks (ISCON), Mathura, India, 1–2 March 2014. [Google Scholar]
  20. Cao, Y.; Ding, Z.; Xue, F.; Rong, X. An improved twin support vector machine based on multi-objective cuckoo search for software defect prediction. Int. J. Bio-Inspired Comput. 2018, 11, 282–291. [Google Scholar] [CrossRef]
  21. Tomar, D.; Agarwal, S. A Multilabel Approach Using Binary Relevance and One-versus-Rest Least Squares Twin Support Vector Machine for Scene Classification. In Proceedings of the Second International Conference on Computational Intelligence & Communication Technology (CICT), Ghaziabad, India, 12–13 February 2016. [Google Scholar]
  22. Gu, Z.; Zhang, Z.; Sun, J.; Li, B. Robust image recognition by L1-norm twin-projection support vector machine. Neurocomputing 2017, 223, 1–11. [Google Scholar] [CrossRef]
  23. Cong, H.; Yang, C.; Pu, X. Efficient Speaker Recognition based on Multi-class Twin Support Vector Machines and GMMs. In Proceedings of the IEEE Conference on Robotics, Automation and Mechatronics, Chengdu, China, 21–24 September 2008. [Google Scholar]
  24. Cumani, S.; Laface, P. Large-Scale Training of Pairwise Support Vector Machines for Speaker Recognition. IEEE/ACM Trans. Audio Speech Lang. Process. 2014, 22, 1590–1600. [Google Scholar] [CrossRef] [Green Version]
  25. Nasiri, J.A.; Charkari, N.M.; Mozafari, K. Energy-based model of least squares twin Support Vector Machines for human action recognition. Signal Process. 2014, 104, 248–257. [Google Scholar] [CrossRef]
  26. Sadewo, W.; Rustam, Z.; Hamidah, H.; Chusmarsyah, A.R. Pancreatic Cancer Early Detection Using Twin Support Vector Machine Based on Kernel. Symmetry 2020, 12, 667. [Google Scholar] [CrossRef] [Green Version]
  27. Peng, X. TPMSVM: A novel twin parametric-margin support vector machine for pattern recognition. Pattern Recognit. 2011, 44, 2678–2692. [Google Scholar] [CrossRef]
  28. Shao, Y.; Zhang, C.; Wang, X.; Deng, N. Improvements on twin support vector machines. IEEE Trans. Neural Netw. 2011, 22, 962–968. [Google Scholar] [CrossRef]
  29. Shao, Y.; Chen, W.; Zhang, C.; Wang, X.; Deng, N. An efficient weighted lagrangian twin support vector machine for imbalanced data classification. Pattern Recognit. 2014, 47, 3158–3167. [Google Scholar] [CrossRef]
  30. Kumar, M.A.; Gopal, M. Application of smoothing technique on twin support vector machines. Pattern Recognit. Lett. 2008, 29, 1842–1848. [Google Scholar] [CrossRef]
  31. Kumar, M.A.; Khemchandani, R.; Gopal, M.; Chandra, S. Knowledge based least squares twin support vector machines. Inform. Sci. 2010, 180, 4606–4618. [Google Scholar] [CrossRef]
  32. Ganaie, M.A.; Tanveer, M. LSTSVM classifier with enhanced features from pre-trained functional link network. Appl. Soft Comput. J. 2020, 93, 106305. [Google Scholar] [CrossRef]
  33. Tian, Y.; Ping, Y. Large-scale linear nonparallel support vector machine solver. Neural Netw. 2014, 50, 166–174. [Google Scholar] [CrossRef]
  34. Tanveer, M.; Tiwari, A.; Choudhary, R.; Jalan, S. Sparse pinball twin support vector machines. Appl. Soft Comput. J. 2019, 78, 164–175. [Google Scholar] [CrossRef]
  35. Lee, Y.J.; Mangasarian, O.L. SSVM: A Smooth Support Vector Machine for Classification. Comput. Optim. Appl. 2001, 20, 5–22. [Google Scholar] [CrossRef]
  36. Wu, Y.; Liu, Y. Robust truncated hinge loss support vector machines. J. Am. Stat. Assoc. 2007, 102, 974–983. [Google Scholar] [CrossRef] [Green Version]
  37. Cao, L.; Shen, H. Imbalanced data classification based on hybrid resampling and twin support vector machine. Comput. Sci. Inf. Syst. 2017, 16, 1–7. [Google Scholar] [CrossRef]
  38. Tomar, D.; Agarwal, S. An effective Weighted Multi-class Least Squares Twin Support Vector Machine for Imbalanced data classification. Int. J. Comput. Intell. Syst. 2015, 8, 761–778. [Google Scholar] [CrossRef] [Green Version]
  39. Huang, X.; Shi, L.; Suykens, J.A.K. Support vector machine classifier with pinball loss. IEEE Trans. Pattern Anal. Mach. Intell. 2014, 36, 984–997. [Google Scholar] [CrossRef]
  40. Rastogi, R.; Pal, A.; Chandra, S. Generalized pinball loss SVMs. Neurocomputing 2018, 322, 151–165. [Google Scholar] [CrossRef]
  41. Mangasarian, O.L. Nonlinear Programming; Society for Industrial and Applied Mathematics: Philadelphia, PA, USA, 1994. [Google Scholar]
  42. Xu, Y.; Yang, Z.; Pan, X. A novel twin support-vector machine with pinball loss. IEEE Trans. Neural Netw. Learn. Syst. 2017, 28, 359–370. [Google Scholar] [CrossRef] [PubMed]
  43. Shwartz, S.; Ben-David, S. Understanding Machine Learning Theory Algorithms; Cambridge University Press: Cambridge, UK, 2014; p. 207. [Google Scholar]
  44. Tikhonov, A.N.; Arsenin, V.Y. Solution of Ill Posed Problems; John Wiley and Sons: Hoboken, NJ, USA, 1977. [Google Scholar]
  45. Khemchandani, R.; Jayadeva; Chandra, S. Optimal kernel selection in twin support vector machines. Optim. Lett. 2009, 3, 77–88. [Google Scholar] [CrossRef]
  46. Dua, D.; Taniskidou, E.K. UCI Machine Learning Repository; University of California, School ofInformation and Computer Science: Irvine, CA, USA, 2019; Available online: http://archive.ics.uci.edu/ml (accessed on 24 September 2018).
  47. Garcı, V.; Sanche, J.S.; Mollineda, R.A. On the effectiveness of preprocessing methods when dealing with different levels of class imbalance. Knowl. Based Syst. 2012, 25, 13–21. [Google Scholar] [CrossRef]
  48. Hsu, C.-W.; Chang, C.-C.; Lina, C.-J. A Practical Guide to Support Vector Classification. Nat. Taiwan Univ. Taipei Taiwa 2012, 25, 1–12. [Google Scholar]
  49. Hamid, N.A.; Sjarif, N.N.A. Handwritten Recognition Using SVM, KNN and Neural Network. arXiv 2017, arXiv:1702.00723. [Google Scholar]
  50. Agarap, A.F.M. An Architecture Combining Convolutional Neural Network (CNN) and Support Vector Machine (SVM) for Image Classification. arXiv 2019, arXiv:1712.03541v2. [Google Scholar]
  51. Ahlawata, S.; Choudhary, A. Hybrid CNN-SVM Classifier for Handwritten Digit Recognition. Procedia Comput. Sci. 2020, 167, 2554–2560. [Google Scholar] [CrossRef]
  52. Aliab, A.A.A.; Mallaiah, S. Intelligent handwritten recognition using hybrid CNN architectures based-SVM classifier with dropout. Comput. Inf. Sci. 2021. [Google Scholar] [CrossRef]
  53. Remaida, A.; Moumen, A.; Idrissi, Y.; El, B.; Sabri, Z. Handwriting Recognition with Artificial Neural Networks a Decade Literature Review. In Proceedings of the 3rd International Conference on Networking, Information Systems & Security, Marrakech, Morocco, 31 March–2 April 2020; pp. 1–5. [Google Scholar]
  54. Aqab, S.; Tariq, M.U. Handwriting Recognition using Artificial Intelligence Neural Network and Image Processing. Int. J. Adv. Comput. Sci. Appl. 2020, 11, 137–146. [Google Scholar] [CrossRef]
  55. Mawaddah, A.H.; Sari, C.A.; Setiadi, D.R.I.M.; Rachmawanto, E.H. Handwriting Recognition of Hiragana Characters using Convolutional Neural Network. In Proceedings of the International Seminar on Application for Technology of Information and Communication (iSemantic), Semarang, Indonesia, 19–20 September 2020. [Google Scholar]
  56. Altwaijry, N.; Al-Turaiki, I. Arabic handwriting recognition system using convolutional neural network. Neural Comput. Appl. 2021, 33, 2249–2261. [Google Scholar] [CrossRef]
  57. Alshazly, H.; Linse, C.; Barth, E.; Martinetz, T. Handcrafted versus CNN Features for Ear Recognition. Symmetry 2019, 11, 1493. [Google Scholar] [CrossRef] [Green Version]
  58. Xin, Q.; Hu, S.; Liu, S.; Ma, X.; Lv, H.; Zhang, Y.D. Epilepsy EEG classification based on convolution support vector machine. J. Med. Imaging Health Inf. 2021, 11, 25–32. [Google Scholar] [CrossRef]
  59. Garcia, S.; Fernandez, A.; Luengo, J.; Herrera, F. Advanced non-parametric tests for multiple comparisons in the design of experiments in computational intelligence and data mining experimental analysis of power. Inf. Sci. 2010, 180, 2044–2064. [Google Scholar] [CrossRef]
Figure 1. The workflow of the improve disadvantages of SVM.
Figure 1. The workflow of the improve disadvantages of SVM.
Symmetry 14 00289 g001
Figure 2. On the 2D synthetic data, a bar graph depicting the accuracies of four algorithms.
Figure 2. On the 2D synthetic data, a bar graph depicting the accuracies of four algorithms.
Symmetry 14 00289 g002
Figure 3. These illustrations show the noise-insensitive features of the material. We have noise samples ranging from r = 0 % to r = 20 % . The slopes of the separating hyperplanes are indicated in brackets in the legend of each figure when we have (a) r = 0 % (free noise); (b) r = 5 ; (c) r = 10 % ; and (d) r = 20 % .
Figure 3. These illustrations show the noise-insensitive features of the material. We have noise samples ranging from r = 0 % to r = 20 % . The slopes of the separating hyperplanes are indicated in brackets in the legend of each figure when we have (a) r = 0 % (free noise); (b) r = 5 ; (c) r = 10 % ; and (d) r = 20 % .
Symmetry 14 00289 g003
Figure 4. Samples from MNIST dataset.
Figure 4. Samples from MNIST dataset.
Symmetry 14 00289 g004
Figure 5. Architecture of the CNN.
Figure 5. Architecture of the CNN.
Symmetry 14 00289 g005
Figure 6. TSVM, Pin-SVM, IPin-TSVM, and GPin-TSVM accuracy and standard deviations on the MNIST dataset.
Figure 6. TSVM, Pin-SVM, IPin-TSVM, and GPin-TSVM accuracy and standard deviations on the MNIST dataset.
Symmetry 14 00289 g006
Table 1. UCI datasets are described in detail.
Table 1. UCI datasets are described in detail.
Datasets#Features#SamplesIR
Breast101161.23
Planning relax121822.5
Ionosphere333511.79
Heart-Statlog132701.25
Heart-C133031.19
Spect222673.85
Saheart94621.89
WDBC305691.68
Pima-Indian87681.86
Australian146901.25
Table 2. On UCI datasets, 10-fold cross validation using the linear kernel yielded the mean accuracy (%) and standard deviation.
Table 2. On UCI datasets, 10-fold cross validation using the linear kernel yielded the mean accuracy (%) and standard deviation.
DatasetsrExisting AlgorithmProposed Algorithm
TSVMPin-TSVMIPin-TSVMGPin-TSVM
Breast069.02 ± 8.6169.85 ± 7.9272.73 ± 16.7971.67 ± 15.53
0.0569.02 ± 8.4669.02 ± 8.6172.73 ± 16.7971.74 ± 15.45
0.168.03 ± 8.0968.26 ± 11.8673.56 ± 14.4473.41 ± 14.32
0.269.85 ± 10.0870.08 ± 13.1571.82 ± 15.8271.82 ± 14.56
Planning relax071.61 ± 13.7871.61 ± 13.7871.61 ± 13.7871.61 ± 13.78
0.0571.61 ± 13.7871.61 ± 13.7871.61 ± 13.7871.61 ± 13.78
0.171.61 ± 13.7871.05 ± 13.4171.61 ± 13.7872.16 ± 13.68
0.271.61 ± 13.7872.69 ± 13.9571.61 ± 13.7871.61 ± 13.78
Australian086.09 ± 3.5686.38 ± 4.4586.96 ± 3.7286.81 ± 3.69
0.0585.36 ± 4.0286.23 ± 4.5085.80 ± 4.5787.39 ± 3.43
0.185.22 ± 3.7185.51 ± 3.6785.36 ± 5.1688.26 ± 4.02
0.283.91 ± 4.9185.22 ± 4.0484.93 ± 4.5586.81 ± 3.69
Heart-Statlog084.07 ± 8.1283.33 ± 7.4583.70 ± 9.4083.33 ± 6.47
0.0583.33 ± 8.3283.70 ± 8.1583.33 ± 9.6984.44 ± 6.99
0.183.70 ± 8.1582.59 ± 7.9583.70 ± 9.5484.07 ± 7.04
0.284.44 ± 8.0883.33 ± 7.0881.11 ± 9.8683.70 ± 8.80
Saheart071.64 ± 5.0171.01 ± 7.4072.08 ± 5.3472.93 ± 6.37
0.0571.63 ± 5.5070.58 ± 7.5171.86 ± 4.4372.71 ± 6.59
0.172.51 ± 5.8470.37 ± 7.2571.86 ± 5.9072.06 ± 5.94
0.271.21 ± 5.5869.93 ± 6.0071.00 ± 5.2871.64 ± 5.39
WDBC095.43 ± 1.6195.96 ± 1.3796.31 ± 2.5497.19 ± 1.61
0.0594.20 ± 3.3494.03 ± 3.1695.08 ± 3.0296.13 ± 2.04
0.193.32 ± 2.1992.97 ± 3.6893.50 ± 3.0595.60 ± 1.97
0.293.32 ± 2.1992.44 ± 1.9393.32 ± 2.5794.55 ± 1.46
Pima076.83 ± 3.7176.70 ± 3.3676.57 ± 4.1277.22 ± 4.07
0.0576.70 ± 3.7476.31 ± 3.5376.44 ± 3.6676.96 ± 3.81
0.176.44 ± 4.2977.22 ± 3.3077.09 ± 3.9676.44 ± 3.87
0.276.18 ± 3.6276.83 ± 3.2377.22 ± 4.0376.18 ± 3.73
Ionosphere090.60 ± 3.6191.46 ± 3.1090.90 ± 5.2092.31 ± 2.87
0.0587.47 ± 6.2890.31 ± 2.9288.89 ± 4.5190.88 ± 3.34
0.187.44 ± 5.9086.88 ± 5.9086.62 ± 5.8389.18 ± 5.05
0.284.34 ± 5.8585.47 ± 5.6484.91 ± 4.3887.16 ± 5.17
Table 3. On UCI datasets, 10-fold cross validation using the RBF kernel yielded the mean accuracy (%) and standard deviation.
Table 3. On UCI datasets, 10-fold cross validation using the RBF kernel yielded the mean accuracy (%) and standard deviation.
DatasetsrExisting AlgorithmProposed Algorithm
TSVMPin-TSVMIPin-TSVMGPin-TSVM
Breat074.24 ± 11.2374.24 ± 11.2378.71 ± 12.4774.24 ± 12.16
0.0574.24 ± 11.2374.24 ± 11.2380.38 ± 10.6174.32 ± 13.35
0.173.33 ± 12.3674.32 ± 10.3480.38 ± 11.2573.41 ± 11.79
0.273.41 ± 13.1573.33 ± 12.3677.80 ± 12.0975.15 ± 13.63
Spect083.50 ± 6.1584.25 ± 6.3184.23 ± 7.0084.63 ± 7.05
0.0583.12 ± 5.6983.49 ± 6.6084.62 ± 6.4684.63 ± 6.43
0.182.75 ± 5.6983.15 ± 7.3283.87 ± 6.9784.64 ± 7.72
0.282.02 ± 5.2282.01 ± 6.9683.12 ± 7.0583.29 ± 7.43
Australian086.81 ± 2.9386.67 ± 3.6086.96 ± 3.9486.67 ± 4.19
0.0582.46 ± 4.0284.78 ± 3.9086.81 ± 4.2787.39 ± 3.11
0.182.61 ± 2.5982.90 ± 4.9386.38 ± 3.3286.52 ± 3.61
0.282.32 ± 3.4282.46 ± 4.7485.51 ± 2.6785.65 ± 3.63
Heart-Statlog084.44 ± 7.7384.81 ± 7.4983.33 ± 8.3284.81 ± 7.49
0.0584.07 ± 7.6084.44 ± 8.0884.81 ± 8.3684.81 ± 7.49
0.184.07 ± 7.9584.44 ± 6.9983.70 ± 8.3184.44 ± 7.91
0.284.07 ± 7.2382.59 ± 7.6084.07 ± 7.4284.07 ± 6.84
Heart-C082.85 ± 4.5282.23 ± 8.9482.85 ± 4.7982.90 ± 8.36
0.0582.19 ± 5.0481.25 ± 8.3081.87 ± 4.0882.54 ± 7.08
0.181.56 ± 7.0480.89 ± 6.7781.89 ± 6.0982.53 ± 4.59
0.280.23 ± 5.9580.81 ± 5.7980.23 ± 4.5582.18 ± 5.46
WDBC097.54 ± 1.1797.71 ± 1.5897.89 ± 1.3297.89 ± 1.32
0.0595.61 ± 2.5195.79 ± 2.2495.61 ± 1.9695.60 ± 2.26
0.195.08 ± 3.1295.08 ± 2.9195.78 ± 2.7495.08 ± 2.58
0.293.49 ± 2.7593.85 ± 1.7994.03 ± 2.2494.20 ± 2.72
Ionosphere096.02 ± 2.5895.17 ± 3.6095.16 ± 2.2395.15 ± 4.05
0.0595.17 ± 3.1294.60 ± 3.9194.87 ± 2.4994.59 ± 3.24
0.194.60 ± 3.2294.60 ± 2.6794.87 ± 3.0794.31 ± 3.36
0.293.46 ± 3.0992.60 ± 2.9192.59 ± 4.0993.48 ± 4.44
Pima077.09 ± 3.3176.96 ± 3.2077.35 ± 3.7177.48 ± 3.33
0.0576.96 ± 3.4576.83 ± 3.3476.70 ± 3.0176.96 ± 3.83
0.176.57 ± 2.5175.66 ± 3.1475.92 ± 3.4276.43 ± 2.68
0.275.53 ± 1.6576.18 ± 3.5176.05 ± 2.8276.57 ± 3.16
Table 4. The optimal parameters of Table 2.
Table 4. The optimal parameters of Table 2.
DatasetsTSVMPin-TSVMIPin-TSVMGPin-TSVM
c 1 , c 2 c 1 , c 2 , τ c 1 , c 2 , τ , ϵ c 1 , c 2 , τ 1 , τ 2 ,
ϵ 1 , ϵ 2
Breast0.01, 0.010.01, 0.01, 110, 0.01, 1, 10.01, 0.01, 0.75, 0.5,
0.1, 0.1
Planning relax0.1, 10.1, 1, 0.10.1, 1, 0.1, 0.10.1, 0.1, 1, 0.1,
0.1, 0.1
Australian1, 0.10.1, 0.1, 0.51, 0.1, 0.1, 11, 0.1, 1, 0.5,
0.1, 0.1
Heart-Statlog0.1, 0.11, 1, 11, 10, 0.5, 11, 10, 1, 0.5,
0.1, 0.5
Saheart0.1, 0.11, 1, 11, 1, 1, 0.11, 1, 1.5, 0.5,
0.5, 0.1
WDBC0.01, 0.010.1, 0.1, 0.11, 10, 0.1, 0.50.01, 0.01, 2, 1,
0.5, 0.1
Pima0.1, 0.10.1, 0.1, 1.51, 1, 1, 0.11, 1, 1, 0.5,
0.1, 0.1
Ionosphere0.01, 0.010.1, 0.1, 0.11, 10, 0.1, 0.51, 10, 1, 0.1,
0.5, 0.5
Table 5. The optimal parameters of Table 3.
Table 5. The optimal parameters of Table 3.
DatasetsTSVMPin-TSVMIPin-TSVMGPin-TSVM
c 1 , c 2 , γ c 1 , c 2 , τ , γ c 1 , c 2 , τ ,
ϵ , γ
c 1 , c 2 , τ 1 , τ 2 ,
ϵ 1 , ϵ 2 , γ
Breast0.01, 0.01, 0.10.01, 0.01, 0.1, 0.10.01, 0.01, 0.1,
1, 0.1
0.1, 0.1, 1, 1,
0.5, 0.5, 0.01
Spect0.1, 0.1, 0.010.1, 0.1, 0.75, 0.010.1, 0.1, 1,
0.5, 0.01
0.1, 0.1, 1, 1,
0.75, 0.75, 0.01
Australian10, 10, 0.0110, 10, 1, 0.0110, 10, 1,
0.1, 0.01
0.01, 0.01, 1, 1,
0.5, 0.5, 0.01
Heart-Statlog0.1, 0.1, 0.011, 1, 1, 0.011, 10, 0.5,
1, 0.01
1, 10, 1, 0.5,
0.1, 0.5, 0.01
Heart-C0.01, 0.01, 0.10.01, 0.01, 0.1, 0.10.01, 0.01, 0.5
0.1, 0.1
0.01, 0.01, 0.5, 0.5
0.1, 0.1, 0.1
WDBC0.1, 0.1, 0.010.1, 0.1, 0.5, 0.011, 1, 1,
0.1, 0.01
0.01, 0.01, 2.5, 2.5,
0.1, 0.1, 0.01
Ionosphere0.1, 1, 0.10.1, 1, 0.5, 0.11, 1, 0.5,
0.1, 0.1
0.1, 0.1, 1, 1,
0.5, 0.5, 0.01
Pima0.1, 0.1, 0.010.1, 0.1, 1.5, 0.10.1, 0.1, 1,
0.5, 0.1
0.1, 0.1, 1, 1,
0.1, 0.1, 0.01
Table 6. Sparsity for UCI datasets employing linear kernel with τ 1 = τ 2 and ϵ 1 = ϵ 2 .
Table 6. Sparsity for UCI datasets employing linear kernel with τ 1 = τ 2 and ϵ 1 = ϵ 2 .
Datasets ϵ 1 TSVMGPTSVM
τ 1 = 0 τ 1 = 0 . 5
Heart-Statlog011213392113
0.05 7885
0.1 6668
0.2 5056
0.3 4646
0.4 3331
Australian0257255254232
0.05 128151
0.1 124136
0.2 113114
0.3 10489
0.4 7269
Breast059475148
0.05 4546
0.1 4044
0.2 3941
0.3 3032
0.4 2423
WDBC0339196234150
0.05 136125
0.1 93103
0.2 6966
0.3 5142
0.4 3026
Ionosphere010819994151
0.05 8981
0.1 7669
0.2 6158
0.3 5146
0.4 4131
Table 7. Sparsity for UCI datasets employing RBF kernel with τ 1 = τ 2 and ϵ 1 = ϵ 2 .
Table 7. Sparsity for UCI datasets employing RBF kernel with τ 1 = τ 2 and ϵ 1 = ϵ 2 .
Datasets ϵ 1 TSVMGPTSVM
τ 1 = 0 τ 1 = 0 . 5
Heart-C0138165138165
0.05 113115
0.1 7883
0.2 6265
0.3 5352
0.4 3838
Spect05521255212
0.05 55121
0.1 5590
0.2 5464
0.3 5056
0.4 3647
Australian0383307383307
0.05 134155
0.1 121138
0.2 114117
0.3 10095
0.4 7167
Ionosphere0126225126225
0.05 9161
0.1 7350
0.2 4141
0.3 3328
0.4 2421
Breast064526452
0.05 5552
0.1 5049
0.2 4244
0.3 3337
0.4 2323
Table 8. The optimal parameters of the result in Figure 6.
Table 8. The optimal parameters of the result in Figure 6.
DatasetsTSVMPin-TSVMIPin-TSVMGPin-TSVM
c 1 , c 2 c 1 , c 2 , τ c 1 , c 2 , τ , ϵ c 1 , c 2 , τ 1 , τ 2 ,
ϵ 1 , ϵ 2
0 vs. 80.1, 0.10.1, 0.1, 0.50.01, 0.01, 1, 10.01, 0.01, 1, 1,
0.1, 0.1
2 vs. 40.1, 0.10.1, 0.1, 0.50.1, 0.1, 1, 0.50.1, 0.1, 1, 1,
0.5, 0.5
1 vs. 71, 11, 1, 0.51, 1, 0.1, 11, 0.1, 1, 0.5,
0.1, 0.1
5 vs. 80.1, 0.11, 1, 11, 1, 0.5, 10.1, 0.1, 1, 0.5,
0.1, 0.5
Table 9. Average rank of different algorithms on four pairs of handwritten digits.
Table 9. Average rank of different algorithms on four pairs of handwritten digits.
DatasetsTSVMPin-TSVMIPin-TSVMGPin-TSVM
0 vs. 84321
2 vs. 43.53.521
1 vs. 7431.51.5
5 vs. 84312
Average Rank3.883.131.631.38
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Panup, W.; Ratipapongton, W.; Wangkeeree, R. A Novel Twin Support Vector Machine with Generalized Pinball Loss Function for Pattern Classification. Symmetry 2022, 14, 289. https://doi.org/10.3390/sym14020289

AMA Style

Panup W, Ratipapongton W, Wangkeeree R. A Novel Twin Support Vector Machine with Generalized Pinball Loss Function for Pattern Classification. Symmetry. 2022; 14(2):289. https://doi.org/10.3390/sym14020289

Chicago/Turabian Style

Panup, Wanida, Wachirapong Ratipapongton, and Rabian Wangkeeree. 2022. "A Novel Twin Support Vector Machine with Generalized Pinball Loss Function for Pattern Classification" Symmetry 14, no. 2: 289. https://doi.org/10.3390/sym14020289

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop