Next Article in Journal
Short-Term Load Forecasting Based on the Transformer Model
Previous Article in Journal
Application of Machine Learning Techniques to Predict the Price of Pre-Owned Cars in Bangladesh
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Multiclass Nonparallel Parametric-Margin Support Vector Machine

1
Zhijiang College, Zhejiang University of Technology, Shaoxing 312030, China
2
College of Computer Science and Technology, Zhejiang University of Technology, Hangzhou 310024, China
3
Management School, Hainan University, Haikou 570228, China
*
Author to whom correspondence should be addressed.
Submission received: 15 November 2021 / Revised: 7 December 2021 / Accepted: 8 December 2021 / Published: 10 December 2021

Abstract

:
The twin parametric-margin support vector machine (TPMSVM) is an excellent kernel-based nonparallel classifier. However, TPMSVM was originally designed for binary classification, which is unsuitable for real-world multiclass applications. Therefore, this paper extends TPMSVM for multiclass classification and proposes a novel K multiclass nonparallel parametric-margin support vector machine (MNP-KSVC). Specifically, our MNP-KSVC enjoys the following characteristics. (1) Under the “one-versus-one-versus-rest” multiclass framework, MNP-KSVC encodes the complicated multiclass learning task into a series of subproblems with the ternary output { 1 , 0 , + 1 } . In contrast to the “one-versus-one” or “one-versus-rest” strategy, each subproblem not only focuses on separating the two selected class instances but also considers the side information of the remaining class instances. (2) MNP-KSVC aims to find a pair of nonparallel parametric-margin hyperplanes for each subproblem. As a result, these hyperplanes are closer to their corresponding class and at least one distance away from the other class. At the same time, they attempt to bound the remaining class instances into an insensitive region. (3) MNP-KSVC utilizes a hybrid classification and regression loss joined with the regularization to formulate its optimization model. Then, the optimal solutions are derived from the corresponding dual problems. Finally, we conduct numerical experiments to compare the proposed method with four state-of-the-art multiclass models: Multi-SVM, MBSVM, MTPMSVM, and Twin-KSVC. Experimental results demonstrate the feasibility and effectiveness of MNP-KSVC in terms of multiclass accuracy and learning time.

1. Introduction

Data mining has become an essential tool for integrating information technology and industrialization due to the growing size of available databases [1]. One of the main applications in data mining is supervised classification. This aims to assign a label to an unseen instance that is as correct as possible from a given set of classes based on the learning model. Recently, support vector machine (SVM) [2,3] has been a preeminent maximum-margin learning paradigm for data classification. The basic idea of SVM is to find an optimal decision boundary via maximizing the margin between two parallel support hyperplanes. Compared with the neural network, SVM has the following attractive features [3]: (1) the structural risk minimization principle is implemented in SVM to control the upper bound of the generalization error, leading to an excellent generalization ability; (2) the global optimum can be achieved by optimizing a quadratic programming problem (QPP). Furthermore, kernel techniques enable SVM to handle complicated nonlinear learning tasks effectively. During recent decades, SVM has been successfully applied in a wide variety of fields ranging from scene classification [4], fault diagnosis [5,6], EEG classification [7,8], pathological diagnosis [9,10], and bioinformatics [8] to power applications [11].
One limitation in the classical SVM is the strictly parallel requirements for support vector hyperplanes. Namely, parallel hyperplanes are challenging to use to capture a data structure with a cross-plane distribution [12,13], such as “XOR” problems. To alleviate the above issue, the nonparallel SVM models have been proposed in the literature [13,14,15,16,17] during the past years. This approach relaxes the parallel requirement in SVM and seeks nonparallel hyperplanes for different classes. The pioneering work is the generalized eigenvalue proximal SVM (GEPSVM) proposed by Mangasarian and Wild [12], which attempts to find a pair of nonparallel hyperplanes via solving eigenvalue problems (EPs). Subsequently, Jayadeva et al. [13] proposed a novel QPP-type nonparallel model for classification, named TWSVM. The idea of TWSVM is to generate two nonparallel hyperplanes such that each hyperplane is closer to one of the two classes and is at least one apart from the other class. Compared with the classical SVM, the nonparallel SVM models (GEPSVM and TWSVM) have lower computational complexity and better generalization ability. Therefore, in the last few years, they have been studied extensively and developed rapidly, including a least squares version of TWSVM (LSTSVM) [16], structural risk minimization version of TWSVM (TBSVM) [17], ν -PTSVM [18], nonparallel SVM (NPSVM) [19,20], nonparallel projection SVM (NPrSVM) [21], and so on [21,22,23,24,25,26,27,28].
The above nonparallel models were mainly proposed for binary classification problems. However, most real-world applications [29,30,31,32] are related to multiclass classifications such as disease diagnosis, fault detection, image recognition, and text categorization. Therefore, many researchers are interested in extending SVM models from binary to multiclass classification. Generally, the decomposition procedure has been considered to be an effective way to achieve multiclass extensions. Yang et al. [33] proposed a multiple birth SVM (MBSVM) for multiclass classification based on the “one-versus-rest” strategy, which is the first multiclass extension of the nonparallel SVM model. Angulo et al. [34] proposed an “one-versus-one-versus-rest” multiclass framework. In contrast to the “one-versus-one” strategy, it constructs k ( k 1 ) / 2 binary classifiers with all data points, which can avoid the risk of information loss and class distortion problems. Following this framework, Xu et al. [35] proposed a multiclass extension of TWSVM, termed Twin-KSVC. Results show that Twin-KSVC has a better generalization ability than MBSVM in most cases. Nasiri et al. [36] formulated Twin-KSVC in the least-squared sense to boost learning efficiency and further presented the LST-KSVC model. Lima et al. [32] proposed an improvement on LST-KSVC (ILST-KSVC) with regularization to implement the structural risk minimization principle.
As a successful extension of SVM, the twin parametric-margin support vector machine (TPMSVM) [15] was proposed to pursue a pair of parametric-margin nonparallel hyperplanes. Unlike GEPSVM and TWSVM, each hyperplane in TPMSVM aims to be closer to its class and far away from the other class. The parametric-margin mechanism enables TPMSVM to be suitable for many cases and results in better generalization performance. However, TPMSVM can only deal with binary classification learning tasks. The above motivates us to propose a novel K multiclass nonparallel parametric-margin support vector machine, termed MNP-KSVC. The proposed MNP-KSVC is endowed with the following attractive advantages:
  • The proposed MNP-KSVC encodes the K multiclass learning task into a series of “one-versus-one-versus-rest” subproblems with all the training instances. Then, it encodes the outputs of subproblems with the ternary output { 1 , 0 , + 1 } , which helps to deal with imbalanced cases.
  • For each subproblem, MNP-KSVC aims to find a pair of nonparallel parametric-margin to separate the two selected classes together with the remaining classes. Unlike TMPSVM, each parametric-margin hyperplane is closer to its class and at least at a distance away from the other class, meanwhile mapping the remaining instances into a region.
  • To implement the empirical risks, MNP-KSVC considers the hybrid classification and regression loss. The Hinge loss is utilized to penalize the errors of the focused two class instances and the ε -intensive loss for the remaining class instances.
  • Extensive numerical experiments are performed on several multiclass UCI benchmark datasets, and their results are compared with four models (Multi-SVM, MBSVM, MTPMSVM, and Twin-KSVC). The comparative results indicate the effectiveness and feasibility of the proposed MNP-KSVC for multiclass classification.
The remainder of this paper is organized as follows. Section 2 briefly introduces notations and related works. Section 3 proposes our MNP-KSVC. The model optimization is also discussed in Section 3. The nonlinear version of MNP-KSVC is extended in Section 4. Experimental results are described in Section 5 and Section 6 presents a discussion and future work.

2. Preliminaries

In this section, we first describe the notations used throughout the paper. Then, we briefly revisit the nonparallel classifier PTSVM and its variants.

2.1. Notations

In this paper, scalars are denoted by lowercase italic letters, vectors by lowercase bold face letters, and matrices by uppercase letters. All vectors are column vectors unless transformed to row vectors by a prime superscript ( · ) . A vector of zeros of arbitrary dimensions is represented by 0 . In addition, we denote e as a vector of ones and I as an identity matrix of arbitrary dimensions. Moreover, let · stand for the L 2 -norm.

2.2. TPMSVM

The TPMSVM [15] is originally proposed for binary classification with a heteroscedastic noise structure. It attempts to seek two nonparallel parametric-margin hyperplanes via the following QPPs:
min w 1 , b 1 1 2 w 1 2 + ν 1 i I + ξ 1 i + c 1 j I η 1 j , s . t . w 1 x i + b 1 ξ 1 i , ξ 1 i 0 , w 1 x j + b 1 = η 1 j ,
and
min w 2 , b 2 1 2 w 2 2 + ν 2 i I ξ 2 i c 2 j I + η 2 j , s . t . w 2 x i + b 2 ξ 1 i , ξ 2 i 0 , w 2 x j + b 2 = η 2 j ,
where ν 1 , ν 2 , c 1 , c 2 are positive parameters, and the decision hyperplane is half of the sum of the two hyperplanes. To obtain the solutions of problems (1) and (2), one needs to resort to the dual problems
min α 1 1 2 i I + j I + α 1 i 2 x i x j ν 1 i I + j I α 1 i x i x j , s . t . i I + α 1 i = ν 1 , 0 α 1 i c 1
and
min α 2 1 2 i I j I α 2 i 2 x i x j ν 2 i I j I + α 2 i x i x j , s . t . i I α 2 i = ν 2 , 0 α 2 i c 2 .
Then, the solutions ( w 1 , b 1 ) and ( w 2 , b 2 ) of dual problems (3) and (4) can be calculated according to the Karush–Kuhn–Tucker (KKT) conditions [3],
w 1 = i I + α 1 i x i ν 1 j I x j and w 2 = i I α 2 i x i + ν 2 j I + x j ,
b 1 = 1 | I S V 1 | i I S V 1 w k x i and b 2 = 1 | I S V 2 | i I S V 2 w k x i ,
where I S V 1 and I S V 2 are indices of the support vector set.
Note that TPMSVM can capture more complex heteroscedastic error structures via parametric-margin hyperplanes compared with TWSVM [13]. However, the variable b in TMPSVM is not strictly convex, leading to a lack of a unique solution. Moreover, the optimization problems (1) and (2) are only designed for binary classification tasks, and thus are unsuitable for many real-world multiclass learning applications.

3. The Proposed MNP-KSVC

3.1. Model Formulation

To address the above issues in TPMSVM, this subsection proposes a novel K multiclass nonparallel parametric-margin support vector machine, termed MNP-KSVC. Inspired by the “hybrid classification and regression” learning paradigm [35], MNP-KSVC decomposes the complicated K multiclass learning task into a series of “one-versus-one-versus-rest” subproblems. Each subproblem focuses on the separation of the two selected classes together with the remaining classes. Here, we utilize { 1 , 1 } to represent the label of the two selected classes and 0 to label the rest. Namely, the subproblem is encoded with the ternary output { 1 , 0 , 1 } . The main idea of MNP-KSVC is to find a pair of nonparallel parametric-margin hyperplanes for each subproblem,
f 1 ( x ) = w 1 x + b 1 = 0 and f 2 ( x ) = w 2 x + b 2 = 0 ,
such that each one is approximate to the corresponding class while as far as possible from the other class on one side. Moreover, the remaining classes are restricted in a region between these hyperplanes.
Formally, we use I k to express the set of indices for instances belonging to the k label in the subproblem, where k is in { 1 , 0 , 1 } . Inspired by the TPMSVM [15], our MNP-KSVC considers the following two loss functions for the above ternary output learning problem:
R 1 e m p = ν 1 j I ( w 1 x j + b 1 ) + c 1 i I + max ( 0 , 1 ( w 1 x i + b 1 ) ) + c 2 l I 0 max ( 0 , ε 1 + ( w 1 x l + b 1 ) )
and
R 2 e m p = ν 2 i I ( w 2 x i + b 2 ) + c 3 j I + max ( 0 , 1 + ( w 2 x j + b 2 ) ) + c 4 l I 0 max ( 0 , ε 1 ( w 2 x l + b 2 ) ) ,
where ν 1 , ν 2 , c 1 , c 2 , c 3 , c 4 > 0 are penalty parameters and ε ( 0 , 1 ] is a margin parameter. Introducing the regularization term w 2 2 + b 2 yields the primal problems of MNP-KSVC:
min w 1 , ξ 1 , η 1 1 2 ( w 1 2 + b 1 2 ) + ν 1 j I ( w 1 x j + b 1 ) + c 1 i I + ξ 1 i + c 3 l I 0 η 1 l s . t . ( w 1 x i + b 1 ) 1 ξ 1 j , ξ 1 i 0 , ( w 1 x l + b 1 ) ε 1 η 1 l , η 1 l 0
and
min w 2 , ξ 2 , η 2 1 2 ( w 2 2 + b 2 2 ) ν 2 i I + ( w 2 x i + b 2 ) + c 2 j I ξ 2 i + c 4 l I 0 η 2 l s . t . ( w 2 x j + b 2 ) 1 ξ 2 j , ξ 2 j 0 , ( w 2 x l + b 2 ) ε 1 η 2 l , η 2 l 0 ,
where ξ 1 , η 1 , ξ 2 , η 2 are non-negative slack vectors.
To deliver the mechanism of MNP-KSVC, we now give the following analysis and geometrical explanation for problem (10):
  • The first term is the L 2 -norm of w 1 and b 1 . Minimizing this is done with the aim of regulating the model complexity of MNP-KSVC and avoiding over-fitting. Furthermore, this regularization term makes the QPPs strictly convex, leading to a unique solution.
  • The second term is the sum of the projection value of the −1 labeled instances x j I on f 1 ( x j ) . Optimizing this term leads instances x j I to be as far as possible from the +1 labeled parametric-margin hyperplane f 1 ( x j ) .
  • The third term with the first constraint requires the projection values of the +1 labeled instances x i I + on hyperplane f 1 ( x j ) to be not less than 1. Otherwise, a slack vector ξ 1 i is introduced to measure its error when the constraint is violated.
  • The last term with the second constraint aims for the projection values of the remaining 0 labeled instances x l I 0 on hyperplane f 1 ( x j ) to be not more than 1 ε . Otherwise, a slack variable η 1 l is utilized to measure the corresponding error. Optimizing this is done with the aim of keeping instances x l I 0 at least ε distance from the +1 labeled instances. Moreover, ε controls the margin between “+” and “0” labeled instances.
The geometrical explanation for problem (11) is similar. Let u 1 = [ w 1 b 1 ] , u 2 = [ w 2 b 2 ] , x ˜ = [ x 1 ] . For the sake of simplicity, denote A = { x ˜ } i I + , B = { x ˜ } j I and C = { x ˜ } l I 0 as the instances belonging to + 1 , 1 , and 0 labels, respectively. Then, the matrix formulations of problems (10) and (11) can be expressed as
min u 1 , ξ 1 , η 1 1 2 u 1 2 + ν 1 e B u 1 + c 1 e + ξ 1 + c 3 e 0 η 1 s . t . A u 1 e + ξ 1 , ξ 1 0 , C u 1 ( ε 1 ) e 0 η 1 , η 1 0
and
min u 2 , ξ 2 , η 2 1 2 u 2 2 ν 2 e + A u 2 + c 2 e ξ 2 + c 4 e 0 η 2 s . t . B u 2 e ξ 2 , ξ 2 0 , C u 2 ( ε 1 ) e 0 η 2 , η 2 0 .
In what follows, we discuss the solutions of problems (12) and (13).

3.2. Model Optimization

To obtain solutions to problems (12) and (13), we first derive their dual problems by Theorem 1.
Theorem 1.
Optimization problems
min α 1 , β 1 1 2 [ α 1 β 1 ] A A A C C A C C α 1 β 1 ν 1 e [ B A B C ] + [ e + ( ε 1 ) e 0 ] α 1 β 1 s . t . 0 α 1 c 1 e + , 0 β 1 c 3 e 0
and
min α 2 , β 2 1 2 [ α 2 β 2 ] B B B C C B C C α 2 β 2 ν 2 e + [ A B A C ] + [ e ( ε 1 ) e 0 ] α 2 β 2 s . t . 0 α 2 c 3 e , 0 β 2 c 4 e 0
are the dual problems of (12) and (13), respectively.
Proof of Theorem 1. 
Taking problem (12) as an example, firstly, we construct its Lagrangian function as
L ( Ξ 1 ) = 1 2 u 1 2 + ν 1 e B u 1 + c 1 e + ξ 1 + c 3 e 0 η 1 α 1 ( A u 1 + ξ 1 e + ) β 1 ( C u 1 + η 1 ( ε 1 ) e 0 ) φ 1 ξ 1 γ 1 η 1 ,
where α 1 , β 1 , φ 1 , γ 1 are the non-negative Lagrange multipliers to constraints of problem (12), and Ξ 1 = { u 1 , ξ 1 , η 1 , α 1 , β 1 , φ 1 , γ 1 } . According to KKT conditions [3,37], the Lagrangian function (16) has to be maximized with its dual variables α 1 , β 1 , φ 1 , γ 1 while being minimized with its primal variables u 1 , ξ 1 , η 1 . Differentiate L ( Ξ 1 ) with respect to u 1 , η 1 , ξ 1 ; then, optimal conditions of problem (12) are obtained by
L u 1 = u 1 + ν 1 B e A α 1 + C β 1 = 0 ,
L ξ 1 = c 1 e + α 1 φ 1 = 0 ,
L η 1 = c 3 e 0 β 1 γ 1 = 0 ,
α 1 ( A u 1 + ξ 1 e + ) = 0 ,
β 1 ( C u 1 + η 1 ( ε 1 ) e 0 ) = 0 ,
φ 1 ξ 1 = 0 ,
γ 1 η 1 = 0 .
From (17), we have
u 1 = A α 1 C β 1 ν 1 B e .
Since φ 1 , γ 1 0 , from (18) and (19), we derive
0 α 1 c 1 e + and 0 β 1 c 3 e 0 .
Finally, substituting (24) into the Lagrangian function (16) and using KKT conditions (17)–(23), the dual problem of (12) can be formulated as
min α 1 , β 1 1 2 [ α 1 β 1 ] A A A C C A C C α 1 β 1 ν 1 e [ B A B C ] + [ e + ( ε 1 ) e 0 ] α 1 β 1 s . t . 0 α 1 c 1 e + , 0 β 1 c 3 e 0 .
Similarly, we can derive the dual problem of (13) as problem (15). □
For ease of notation, define q 1 = [ α 1 ; β 1 ] , N 1 = [ A ; C ] , h 1 = [ e + ; ( ε 1 ) e 0 ] , and e q 1 = [ c 1 e + ; c 3 e 0 ] . Then, problem (14) can be succinctly reformulated as
min q 1 1 2 q 1 H 1 q 1 d 1 q 1 s . t . 0 q 1 e q 1
where
H 1 = N 1 N 1 = A A A C C A C C
and
d 1 = ν 1 e B N 1 + h 1 = ν 1 e [ B A B C ] + [ e + ( ε 1 ) e 0 ]
Similarly, define q 2 = [ α 2 ; β 2 ] , N 2 = [ B ; C ] , h 2 = [ e ; ( ε 1 ) e 0 ] , and e q 2 = [ c 2 e ; c 4 e 0 ] for problem (15). Then, it can be reformulated as
min q 2 1 2 q 2 H 2 q 2 d 2 q 2 s . t . 0 q 2 e q 2
where
H 2 = N 2 N 2 = B B B C C B C C
and
d 2 = ν 2 e + A N 2 + h 2 = ν 2 e + [ A B A C ] + [ e ( ε 1 ) e 0 ]
After solving dual problems (27) and (30) by the standard QPP solver, we can obtain the solutions to primal problems (12) and (13) by Proposition 1 according to KKT conditions without proof.
Proposition 1.
Suppose that q 1 = [ α 1 ; β 1 ] and q 2 = [ α 2 ; β 2 ] are solutions to dual problems (27) and (30), respectively. Then, solutions u 1 and u 2 to primal problems (12) and (13) can be formulated by
u 1 = N 1 q 1 ν 1 B e = A α 1 C β 1 ν 1 B e
and
u 2 = N 2 q 2 + ν 2 A e + = B α 2 + C β 2 + ν 2 A e +

3.3. Decision Rule

As mentioned in Section 3.1, our MNP-KSVC decomposes the multiclass learning task into a series of subproblems with the “one-versus-one-versus-rest” strategy. Specifically, we construct K ( K 1 ) / 2 MNP-KSVC classifiers for K-class classification. For each ( k 1 , k 2 ) -pair classifier, we relabel the dataset with ternary outputs { 1 , 0 , + 1 } according to the two selected and the remaning class instances. Namely, we label “+1”, “−1”, and “0” to instances belonging to the k i class, k j class, and all the remaining classes, respectively. Then, we train it on the new relabeled dataset by solving problems (12) and (13).
As for the decision, we predict the label of an unseen instance x on the voting strategy from the ensemble results of K ( K 1 ) / 2 MNP-KSVC classifiers. Namely, we determine the vote via each classifier according to the regions in which x is located. Taking the ( k 1 , k 2 ) -pair classifier as an example, firstly, we reformulate the parametric-margin hyperplanes (7) w.r.t. u = [ w ; b ] and x ˜ = [ x 1 ] as
f 1 ( x ) = w 1 x + b 1 = u 1 x ˜ and f 2 ( x ) = w 2 x + b 2 = u 2 x ˜ ,
If x is located above the “+” hyperplane f 1 ( x ) > 1 ε —i.e., satisfying the condition u 1 x ˜ > 1 ε —we vote it to the k i class. On the other hand, if x is located below the “−” hyperplane f 2 ( x ) < 1 + ε , we vote it to the k j class. Otherwise, it belongs to the remaining classes. In summary, the decision function for the ( k 1 , k 2 ) -pair classifier can be expressed as
g ( x ) = + 1 f 1 ( x ) > 1 ε , 1 f 2 ( x ) < 1 + ε , 0 otherwise .
Finally, the given unseen instance x is assigned to the class label that gets the most votes. In summary, the whole procedure of MNP-KSVC is established in Algorithm 1 with the Figure 1.
Algorithm 1 The procedure of MNP-KSVC
1:
Input dataset T = { ( x i , y i ) | 1 i m } , where x i R n and y i = { 1 , , K } .
2:
Choose parameters ν 1 , ν 2 , c 1 , c 2 , c 3 , c 4 > 0 and ε ( 0 , 1 ] .
Training Procedure:
3:
for k i in range ( 1 , , K )  do
4:
  for k j in range ( k i + 1 , , K )  do
5:
   Relabel “+1”, “−1”, and “0” to instances belonging to k i , k j , and the rest classes.
6:
   Construct A , B , and C for problem (12) and (13) of ( k 1 , k 2 ) -pair classifier.
7:
   Solve the corresponding dual problem (27) and (30) by QPP solver. Then, get
   their solution q 1 = [ α 1 ; β 1 ] and q 2 = [ α 2 ; β 2 ] .
8:
   Build the auxiliary functions for classifier w.r.t. Proposition 1
k 1 - class : f 1 ( x ) = w 1 x + b 1 = q 1 N 1 x ν 1 e B x
   and
k 2 - class : f 2 ( x ) = w 2 x + b 2 = q 2 N 2 x + ν 2 e + A x
9:
  end for
10:
end for
Predicting Procedure:
11:
For an unseen instance x , assign it to class y via the following vote strategy.
12:
Initialize the vote vector v o t e = 0 for K classes.
13:
for k i in range ( 1 , , K )  do
14:
  for k j in range ( k i + 1 , , K )  do
15:
   Compute the decision function g ( x ) for the ( k 1 , k 2 ) -pair classifier w.r.t. (36).
16:
   if g ( x ) = = 1 then
17:
    Update v o t e ( k i ) + = 1
18:
   else if g ( x ) = = 1 then
19:
    Update vote ( k j ) + = 1
20:
   end if
21:
  end for
22:
end for
23:
Finally, assign the most votes of class to x via
label ( x ) arg max k = { 1 , , K } v o t e ( k )

4. Model Extension to the Nonlinear Case

In practice, the linear classifier is sometimes not suitable for many real-world nonlinear learning tasks [18,21,30]. One of the effective solutions is to map linearly non-separable instances into the feature space. Thus, in this section, we focus on the nonlinear extension of MNP-KSVC.
To construct our nonlinear MNP-KSVC, we consider the feature mapping x φ = φ ( x ) : R n H (RKHS, Reproducing Kernel Hilbert Space) with the kernel trick [3]. Define A ϕ = { ϕ ( x ˜ i ) } i I + , B ϕ = { ϕ ( x ˜ j ) } j I , and C ϕ = { ϕ ( x ˜ l ) } j I 0 . Then, the nonlinear MNP-KSVC model optimizes the following two primal QPPs:
min u 1 , ξ 1 , η 1 1 2 u 1 2 + ν 1 e B ϕ u 1 + c 1 e + ξ 1 + c 3 e 0 η 1 s . t . A ϕ u 1 e + ξ 1 , ξ 1 0 , C ϕ u 1 ( ε 1 ) e 0 η 1 , η 1 0
and
min u 2 , ξ 2 , η 2 1 2 u 2 2 ν 2 e + A ϕ u 2 + c 2 e ξ 2 + c 4 e 0 η 2 s . t . B ϕ u 2 e ξ 2 , ξ 2 0 , C ϕ u 2 ( ε 1 ) e 0 η 2 , η 2 0 .
where ξ 1 , η 1 , ξ 2 , η 2 are non-negative slack vectors. Because the formulations of nonlinear problems (40) and (41) are similar to the linear cases (12) and (13), we can obtain their solutions in a similar manner.
In what follows, we define the kernel operation for MNP-KSVC.
Definition 1.
Suppose that K ( · , · ) is an appropriate kernel function; then, the kernel operation in matrix form is defined as
K ( A , B ) = A ϕ , B ϕ ,
whose i j -th element can be computed by
K ( A , B ) i j = ϕ ( x i ) ϕ ( x j ) = K ( x i , x j ) .
Then, we can derive the dual problems of (40) and (41) by Theorem 2.
Theorem 2.
Optimization problems
min α 1 , β 1 1 2 [ α 1 β 1 ] K ( A , A ) K ( A , C ) K ( C , A ) K ( C , C ) α 1 β 1 ν 1 e [ K ( B , A ) K ( B , C ) ] + [ e + ( ε 1 ) e 0 ] α 1 β 1 s . t . 0 α 1 c 1 e + , 0 β 1 c 3 e 0
and
min α 2 , β 2 1 2 [ α 2 β 2 ] K ( B , B ) K ( B , C ) K ( C , B ) K ( C , C ) α 2 β 2 ν 2 e + [ K ( A , B ) K ( A , C ) ] + [ e ( ε 1 ) e 0 ] α 2 β 2 s . t . 0 α 2 c 3 e , 0 β 2 c 4 e 0
are the dual problems of (40) and (41), respectively.
Proof of Theorem 2. 
By introducing non-negative Lagrange multipliers α 1 , β 1 , φ 1 , and γ 1 to constraints of problem (40), its Lagrangian function is built as
L ( Ξ 1 ) = 1 2 u 1 2 + ν 1 e B ϕ u 1 + c 1 e + ξ 1 + c 3 e 0 η 1 α 1 ( A ϕ u 1 + ξ 1 e + ) β 1 ( C ϕ u 1 + η 1 ( ε 1 ) e 0 ) φ 1 ξ 1 γ 1 η 1 ,
where Ξ 1 = { u 1 , ξ 1 , η 1 , α 1 , β 1 , φ 1 , γ 1 } . Differentiate L ( Ξ 1 ) with respect to u 1 , η 1 , ξ 1 ; then, optimal conditions of problem (40) are obtained by
L u 1 = u 1 + ν 1 B ϕ e A ϕ α 1 + C ϕ β 1 = 0 ,
L ξ 1 = c 1 e + α 1 φ 1 = 0 ,
L η 1 = c 3 e 0 β 1 γ 1 = 0 ,
α 1 ( A ϕ u 1 + ξ 1 e + ) = 0 ,
β 1 ( C ϕ u 1 + η 1 ( ε 1 ) e 0 ) = 0 ,
φ 1 ξ 1 = 0 ,
γ 1 η 1 = 0 .
From (47), we have
u 1 = A ϕ α 1 C ϕ β 1 ν 1 B ϕ e .
Since φ 1 , γ 1 0 , from (48) and (49), we derive
0 α 1 c 1 e + and 0 β 1 c 3 e 0 .
Finally, substituting (54) into the Lagrangian function (16) and using KKT conditions (47)–(53), the dual problem of (40) can be formulated as (44). Similarly, we can derive the dual problem of (41) as problem (45). □
The procedure of the nonlinear MNP-KSVC is similar to that of the linear one, but with the following minor modifications in Algorithm 1:
  • In contrast to some existing nonparallel SVMs, we do not need to consider the extra kernel-generated technique since only inner products appear in dual problems (14) and (15) of the linear MNP-KSVC. These dual formulations enable MNP-KSVC to behave consistently in linear and nonlinear cases. Thus, taking appropriate kernel functions instead of inner products in the Hessian matrix of dual problem (14) and (15), i.e.,
    H ϕ 1 = K ( A , A ) K ( A , C ) K ( C , A ) K ( C , C ) and H ϕ 2 = K ( B , B ) K ( B , C ) K ( C , B ) K ( C , C ) ,
    we can obtain the dual formation of the nonlinear MNP-KSVC in (44) and (45).
  • Once we obtain the solution q 1 = [ α 1 ; β 1 ] and q 2 = [ α 2 ; β 2 ] to problems (44) and (45) respectively, the corresponding primal solutions u 1 and u 2 in feature space can be formulated by
    u 1 = A ϕ α 1 C ϕ β 1 ν 1 B ϕ e
    and
    u 2 = B ϕ α 2 + C ϕ β 2 + ν 2 A ϕ e +
  • For an unseen instance x , construct the following decision functions for ( i , j ) -pair nonlinear MNP-KSVC classifier as
    g ϕ ( x ) = + 1 f ϕ 1 ( x ˜ ) > 1 ε , 1 f ϕ 2 ( x ˜ ) < 1 + ε , 0 otherwise ,
    where x ˜ = [ x 1 ] , the auxiliary functions f ϕ 1 ( x ) and f ϕ 2 ( x ) in feature space can be expressed as
    f ϕ 1 ( x ˜ ) ) = K ( u 1 , x ˜ ) = α 1 K ( A , x ˜ ) β 1 K ( C , x ˜ ) ν 1 e K ( B , x ˜ )
    and
    f ϕ 2 ( x ˜ ) ) = K ( u 2 , x ˜ ) = α 2 K ( B , x ˜ ) + β 2 K ( C , x ˜ ) + ν 2 e + K ( A , x ˜ )

5. Numerical Experiments

5.1. Experimental Setting

To demonstrate the validity of MNP-KSVC, we perform extensive experiments on several benchmark datasets that are commonly used for testing machine learning algorithms. In experiments, we focus on comparing MNP-KSVC and four state-of-the-art multiclass models—Multi-SVM, MBSVM, MTPMSVM, and Twin-KSVC—detailed as follows:
  • Multi-SVM [38]: The idea is similar to the “one-versus-all” SVM [3]. However, it generates K binary SVM classifiers by solving the one large dual QPP. That is, the k-th classifier is trained with the k-th class instances encoded with positive labels and the remaining class instances with negative labels. Then, the label of an unseen instance is assigned by the “voting” scheme. The penalty parameter for each classifier in Multi-SVM is c.
  • MBSVM [33]: It is the multiclass extension of the binary TWSVM, which is based on the “one-versus-all” strategy. MBSVM aims to find K nonparallel hyperplanes by solving K QPPs simultaneously. Specifically, the k-th class instances are as far away as the k-th hyperplane while the remaining instances are proximal to the k-th hyperplane. An unseen instance is assigned to the label depending on to which of the K hyperplanes it lies farthest away. The penalty parameter for each classifier in MBSVM is c.
  • MTPMSVM: Inspired by MBSVM [33], we use the “one-versus-all” strategy to implement the multiclass version of TPMSVM [15] as a baseline. In contrast to MBSVM, it aims to find K parametric-margin hyperplanes, such that each hyperplane is closer to its corresponding class instances and as far away from the remaining class instances. The penalty parameters for each classifier in MTPMSVM are ( ν , c ) .
  • Twin-KSVC [35]: It is another novel multiclass extension of TWSVM. Twin-KSVC evaluates all the training instances in a “one-versus-one-versus-rest” structure with the ternary output { 1 , 0 , + 1 } . It aims to find a pair of nonparallel hyperplanes for each of two kinds of samples selected from K classes. The remaining class instances are mapped into a region within these two nonparallel hyperplanes. The penalty parameters for each classifier in Twin-KSVC are ( c 1 , c 2 , c 3 , c 4 , ε ) .
All methods are implemented by MATLAB on a PC with an i7 Intel Core processor with 32 GB RAM. The quadratic programming problems (QPPs) of all the classifiers are solved by the “quadprog” function in MATLAB. Now, we describe the setting of our experiments:
  • Similar to [35,38], we use multiclass accuracy to measure each classifier, defined as
    accuracy = 1 N k = 1 K x I k I g ^ ( x ) = k
    where N is the scalar of the dataset, K is the total classes, g ^ ( x ) is the prediction of the classifier, and I ( · ) is an indicator function, which returns 1 if the class matches and 0 otherwise. Moreover, we adopt training time to represent the learning efficiency.
  • To reduce the complexity of parameter selection for multiclass classifiers, we use the same parameter setting for each learning subproblem. Specifically, we set the same for all classifiers c in MBSVM and MBSVM, ν , c in MTPSVM, c 1 = c 2 , c 3 = c 4 in Twin-KSVC, and ν 1 = ν 2 , c 1 = c 2 , c 3 = c 4 in MNP-KSVC. For the nonlinear case, the RBF kernel K ( x i , x j ) = exp ( x i x j 2 γ ) is considered, where γ > 0 is the kernel parameter.
  • It is usually unknown beforehand which parameters are optimal for classifiers at hand. Thus, we employ the 10-fold cross-validation technique [3] for parameter selection. In detail, each dataset is randomly partitioned into 10 subsets with similar sizes and distributions. Then, the union of 9 subsets is used as the training set, while the remaining 1 is used as the testing set. Furthermore, we apply the grid-based approach [3] to obtain the optimal parameters of each classifier. Namely, the penalty parameters c , c 1 , c 2 , ν , ν 1 and the kernel parameter γ are selected from { 2 i | i = 6 , , 6 } , while the margin parameter ε is chosen from { i | i = 0.1 , 0.2 , , 0.9 } . Once selected, we return them to learn the final decision function.

5.2. Result Comparison and Discussion

For comparison, we consider 10 real-world multiclass datasets from the UCI machine learning repository (the UCI datasets are available at http://archive.ics.uci.edu/ml (accessed on 10 September 2021)), whose statistics are summarized in Table 1. These datasets represent a wide range of domains (include phytology, bioinformatics, pathology, and so on), sizes (from 178 to 2175), features (from 4 to 34), and classes (from 3 to 10). All datasets are normalized before training such that features are transformed into [ 1 , 1 ] . Moreover, we carry out experiments as follows. Firstly, each dataset is divided into 2 subsets: 70% for training and 30% for testing. Then, we train classifiers with 10-fold cross-validation executions. Finally, we predict the testing set with the fine-tuning classifiers. Each experiment is repeated 10 times.
Table 2 and Table 3 contain a summary of learning results for the proposed MNP-KSVC model with other compared methods using linear and nonlinear kernels, respectively. The results on 10 benchmark datasets include the mean and standard of the testing multiclass accuracy (%), whose best performance is highlighted in bold. The comparison results reveal the following:
  • MNP-KSVC yields better performance than other classifiers in terms of accuracy on almost all datasets. This confirms the efficacy of the proposed MNP-KSVC on the multiclass learning tasks.
  • Nonparallel based models (MBSVM, MTPMSVM, Twin-KSVC, and MNP-KSVC) outperform the traditional SVM model. The reason is that SVM simply utilizes the parallel hyperplanes to learn the decision function, leading to less capability to capture the underlying multiclass distributions.
  • MNP-KSVC has a better generalization ability than MBSVM and Twin-KSVC in most cases. For instance, MNP-KSVC obtains a higher accuracy (84.59%) than MBSVM (80.82%) and Twin-KSVC (80.12%) on the Ecoli dataset in the linear case. Similar results can be obtained from the other datasets. Since MBSVM and Twin-KSVC simply implement empirical risk minimization, they are easy to overfit. The regularization terms are considered in our MNP-KSVC, which regulates the model complexity to avoid overfitting.
  • MTPMSVM is another multiclass extension, which is based on the “one-versus-rest” strategy. With the help of the “hybrid classification and regression” learning paradigm, our MNP-KSVC can learn more multiclass discriminate information.
  • Furthermore, we count the number of Superior/Inferior (W/L) instances to the compared classifier on all datasets for both linear and nonlinear cases, listed at the bottom of Table 2 and Table 3. The results indicate that MNP-KSVC achieves the best results against others in terms of both W/L and average accuracy.
To provide more statistical evidence [39,40,41], we have further performed the non-parametric Friedman test to check whether there are significant differences between MNP-KSVC and other compared classifiers. The bottom lines of Table 2 and Table 3 list the average ranks of each classifier calculated by the Friedman test according to multiclass accuracy. Results show that MNP-KSVC has the first rank in linear and nonlinear cases, followed by NPSVM and RPTSVM successively. Now, we calculate the X F 2 value for the Friedman test as
X F 2 = 12 × N k ( k + 1 ) i = 1 k r i 2 k ( k + 1 ) 2 4
where N is the number of datasets, k is the number of classifiers, and r i is the average rank on N datasets for the i-th model. For the linear case, we compute term i = 1 k r i 2 in Table 2 as
i = 1 k r i 2 = 4.6 2 + 2.9 2 + 3.1 2 + 2.7 2 + 1.7 2 49.1989
Then, substituting k = 5 , N = 10 and (64) into (63), we have
X F 2 = 12 × 10 5 ( 5 + 1 ) 49.1989 5 ( 5 + 1 ) 2 4 16.7956
Based on the above Friedman statistic X F 2 = 16.7956 , we calculate the F-distribution statistic F F with ( k 1 , ( k 1 ) ( N 1 ) ) = ( 4 , 36 ) degrees of freedom as
F F = ( N 1 ) × X F 2 N ( k 1 ) X F 2 = 9 × 16.7956 10 × 4 16.7956 6.5143
Moreover, we compute the p-value, which rejects the null hypothesis at the level of significance α = 0.05 . Similarly, we calculate the statistic for the nonlinear case, as summarized in Table 4. The results reject the null hypothesis for both linear and nonlinear cases and reveal the existence of significant differences in the performances of classifiers.
Furthermore, we record the average learning time of each classifier for the above UCI datasets experiments, as shown in Figure 2 and Figure 3. The results show that our MNP-KSVC is faster than Multi-SVM and Twin-KSVC, while slightly slower than MBSVM and MTPMSVM for linear and nonlinear cases. Multi-SVM performs the slowest of all classifiers because Multi-SVM needs to solve larger problems than the nonparallel-based classifiers. Moreover, the Hessian matrix of dual QPPs in MNP-KSVC avoids the time-costly matrix inversion, leading to greater effectiveness than Twin-KSVC. Overall, the above results confirm the feasibility of Twin-KSVC.

6. Discussion and Future Work

This paper proposes a novel K multiclass nonparallel parametric-margin support vector machine, termed MNP-KSVC. Specifically, our MNP-KSVC has the following attractive merits:
  • For the K-class learning task, our MNP-KSVC first transforms the complicated multiclass problem into K ( K 1 ) / 2 subproblems via a “one-versus-one-versus-rest” strategy. Each subproblem focuses on separating the two selected classes and the rest of the classes. That is, we utilize { 1 , 1 } to represent the label of the two selected classes and 0 to label the rest. Unlike the “one-versus-all” strategy used in Multi-SVM, MBSVM, and MTPMSVM, this encoding strategy can alleviate the imbalanced issues that sometimes occur in multiclass learning [32,35].
  • For each subproblem, our MNP-KSVC aims to learn a pair of nonparallel parametric-margin hyperplanes (36) with the ternary encoding { 1 , 0 , + 1 } . These parametric-margin hyperplanes are closer to their corresponding class and at least one distance from the other class. Meanwhile, they restrict the rest of the instances into an insensitive region. A hybrid classification and regression loss joined with the regularization is further utilized to formulate the optimization problems (10) and (11) of MNP-KSVC.
  • Moreover, the nonlinear extension is also presented to deal with the nonlinear multiclass learning tasks. In contrast to MBSVM [33] and Twin-KSVC [35], the linear and nonlinear models in MNP-KSVC are consistent. Applying the linear kernel in the nonlinear problems (44) and (45) results in the same formulations as the original linear problems (14) and (15).
  • Extensive experiments on various datasets demonstrate the effectiveness of the proposed MNP-KSVC compared with Multi-SVM, MBSVM, MTPMSVM, and Twin-KSVC.
There are several interesting directions to research in the future, such as extensions to semi-supervised learning [26,42], multi-label learning [22], and privilege-information learning [43].

Author Contributions

Conceptualization, S.-W.D. and W.-J.C.; Funding acquisition, W.-J.C. and Y.-H.S.; Investigation, M.-C.Z. and P.C.; Methodology, S.-W.D., M.-C.Z. and Y.-H.S.; Project administration, S.-W.D. and W.-J.C.; Supervision, W.-J.C. and Y.-H.S.; Validation, M.-C.Z. and H.-F.S.; Visualization, P.C. and H.-F.S.; Writing—original draft, S.-W.D., P.C. and H.-F.S.; Writing—review and editing, S.-W.D. and W.-J.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China, Grant number: 61603338, 11871183, 61866010, and 11426202; the Natural Science Foundation of Zhejiang Province of China, Grant number: LY21F030013; the Natural Science Foundation of Hainan Province of China, Grant number: 120RC449; the Scientific Research Foundation of Hainan University, Grant number: kyqd(sk)1804.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data sharing not applicable. No new data were created or analyzed in this study.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
QPPQuadratic Problem Programming
KKTKarush–Kuhn–Tucker
SVMSupport Vector Machine
TWSVMTwin Support Vector Machine
Multi-SVMMulticlass Support Vector Machine
MBSVMMultiple Birth Support Vector Machine
MTPMSVMMultiple Twin Parametric Margin Support Vector Machine
Twin-KSVCMulticlass Twin Support Vector Classifier
MNP-KSVCMulticlass Nonparallel Parametric-Margin Support Vector Classifier

References

  1. Han, J.; Kamber, M.; Pei, J. Data Mining: Concepts and Techniques; Morgan Kaufmann: San Francisco, CA, USA, 2012. [Google Scholar]
  2. Vapnik, V. Statistical Learning Theory; Wiley: New York, NY, USA, 1998. [Google Scholar]
  3. Deng, N.; Tian, Y.; Zhang, C. Support Vector Machines: Theory, Algorithms and Extensions; CRC Press: Philadelphia, PA, USA, 2013. [Google Scholar]
  4. Sitaula, C.; Aryal, S.; Xiang, Y.; Basnet, A.; Lu, X. Content and context features for scene image representation. Knowl.-Based Syst. 2021, 232, 107470. [Google Scholar] [CrossRef]
  5. Ma, S.; Cheng, B.; Shang, Z.; Liu, G. Scattering transform and LSPTSVM based fault diagnosis of rotating machinery. Mech. Syst. Signal Process. 2018, 104, 155–170. [Google Scholar] [CrossRef]
  6. Liu, T.; Yan, D.; Wang, R.; Yan, N.; Chen, G. Identification of Fake Stereo Audio Using SVM and CNN. Information 2021, 12, 263. [Google Scholar] [CrossRef]
  7. You, S.D. Classification of Relaxation and Concentration Mental States with EEG. Information 2021, 12, 187. [Google Scholar] [CrossRef]
  8. Kang, J.; Han, X.; Song, J.; Niu, Z.; Li, X. The identification of children with autism spectrum disorder by SVM approach on EEG and eye-tracking data. Comput. Biol. Med. 2020, 120, 103722. [Google Scholar] [CrossRef]
  9. Lazcano, R.; Salvador, R.; Marrero-Martin, M.; Leporati, F.; Juarez, E.; Callico, G.M.; Sanz, C.; Madronal, D.; Florimbi, G.; Sancho, J.; et al. Parallel Implementations Assessment of a Spatial-Spectral Classifier for Hyperspectral Clinical Applications. IEEE Access 2019, 7, 152316–152333. [Google Scholar] [CrossRef]
  10. Fabelo, H.; Ortega, S.; Szolna, A.; Bulters, D.; Pineiro, J.F.; Kabwama, S.; J-O’Shanahan, A.; Bulstrode, H.; Bisshopp, S.; Kiran, B.R.; et al. In-Vivo Hyperspectral Human Brain Image Database for Brain Cancer Detection. IEEE Access 2019, 7, 39098–39116. [Google Scholar] [CrossRef]
  11. Roy, S.D.; Debbarma, S. A novel OC-SVM based ensemble learning framework for attack detection in AGC loop of power systems. Electr. Power Syst. Res. 2022, 202, 107625. [Google Scholar] [CrossRef]
  12. Mangasarian, O.L.; Wild, E.W. Multisurface proximal support vector machine classification via generalized eigenvalues. IEEE Trans. Pattern Anal. Mach. Intell. 2006, 28, 69–74. [Google Scholar] [CrossRef]
  13. Jayadeva; Khemchandani, R.; Chandra, S. Twin Support Vector Machines for Pattern Classification. IEEE Trans. Pattern Anal. Mach. Intell. 2007, 29, 905–910. [Google Scholar] [CrossRef]
  14. Ding, S.; Hua, X. An overview on nonparallel hyperplane support vector machine algorithms. Neural Comput. Appl. 2014, 25, 975–982. [Google Scholar] [CrossRef]
  15. Peng, X. TPMSVM: A novel twin parametric-margin support vector machine for pattern recognition. Pattern Recogn. 2011, 44, 2678–2692. [Google Scholar] [CrossRef]
  16. Arun Kumar, M.; Gopal, M. Least squares twin support vector machines for pattern classification. Expert Syst. Appl. 2009, 36, 7535–7543. [Google Scholar] [CrossRef]
  17. Shao, Y.; Zhang, C.; Wang, X.; Deng, N. Improvements on Twin Support Vector Machines. IEEE Trans. Neural Netw. 2011, 22, 962–968. [Google Scholar] [CrossRef]
  18. Chen, W.J.; Shao, Y.H.; Li, C.N.; Liu, M.Z.; Wang, Z.; Deng, N.Y. ν-projection twin support vector machine for pattern classification. Neurocomputing 2020, 376, 10–24. [Google Scholar] [CrossRef]
  19. Tian, Y.; Qi, Z.; Ju, X.; Shi, Y.; Liu, X. Nonparallel Support Vector Machines for Pattern Classification. IEEE Trans. Cybern. 2014, 44, 1067–1079. [Google Scholar] [CrossRef] [Green Version]
  20. Tian, Y.; Ping, Y. Large-scale linear nonparallel support vector machine solver. Neural Netw. 2014, 50, 166–174. [Google Scholar] [CrossRef] [PubMed]
  21. Chen, W.; Shao, Y.; Li, C.; Wang, Y.; Liu, M.; Wang, Z. NPrSVM: Nonparallel sparse projection support vector machine with efficient algorithm. Appl. Soft Comput. 2020, 90, 106142. [Google Scholar] [CrossRef]
  22. Chen, W.; Shao, Y.; Li, C.; Deng, N. MLTSVM: A novel twin support vector machine to multi-label learning. Pattern Recogn. 2016, 52, 61–74. [Google Scholar] [CrossRef]
  23. Bai, L.; Shao, Y.H.; Wang, Z.; Chen, W.J.; Deng, N.Y. Multiple Flat Projections for Cross-Manifold Clustering. IEEE Trans. Cybern. 2021, 1–15. Available online: https://0-ieeexplore-ieee-org.brum.beds.ac.uk/document/9343292 (accessed on 20 October 2021). [CrossRef]
  24. Hou, Q.; Liu, L.; Zhen, L.; Jing, L. A novel projection nonparallel support vector machine for pattern classification. Eng. Appl. Artif. Intell. 2018, 75, 64–75. [Google Scholar] [CrossRef]
  25. Liu, L.; Chu, M.; Gong, R.; Zhang, L. An Improved Nonparallel Support Vector Machine. IEEE Trans. Neural Netw. Learn. Syst. 2021, 32, 5129–5143. [Google Scholar] [CrossRef]
  26. Chen, W.; Shao, Y.; Deng, N.; Feng, Z. Laplacian least squares twin support vector machine for semi-supervised classification. Neurocomputing 2014, 145, 465–476. [Google Scholar] [CrossRef]
  27. Shao, Y.; Chen, W.; Zhang, J.; Wang, Z.; Deng, N. An efficient weighted Lagrangian twin support vector machine for imbalanced data classification. Pattern Recogn. 2014, 47, 3158–3167. [Google Scholar] [CrossRef]
  28. Chen, W.; Shao, Y.; Ning, H. Laplacian smooth twin support vector machine for semi-supervised classification. Int. J. Mach. Learn. Cybern. 2014, 5, 459–468. [Google Scholar] [CrossRef]
  29. Gao, Z.; Fang, S.C.; Gao, X.; Luo, J.; Medhin, N. A novel kernel-free least squares twin support vector machine for fast and accurate multi-class classification. Knowl.-Based Syst. 2021, 226, 107123. [Google Scholar] [CrossRef]
  30. Ding, S.; Zhao, X.; Zhang, J.; Zhang, X.; Xue, Y. A review on multi-class TWSVM. Artif. Intell. Rev. 2017, 52, 775–801. [Google Scholar] [CrossRef]
  31. Qiang, W.; Zhang, J.; Zhen, L.; Jing, L. Robust weighted linear loss twin multi-class support vector regression for large-scale classification. Signal Process. 2020, 170, 107449. [Google Scholar] [CrossRef]
  32. de Lima, M.D.; Costa, N.L.; Barbosa, R. Improvements on least squares twin multi-class classification support vector machine. Neurocomputing 2018, 313, 196–205. [Google Scholar] [CrossRef]
  33. Yang, Z.; Shao, Y.; Zhang, X. Multiple birth support vector machine for multi-class classification. Neural Comput. Appl. 2013, 22, 153–161. [Google Scholar] [CrossRef]
  34. Angulo, C.; Parra, X.; Català, A. K-SVCR. A support vector machine for multi-class classification. Neurocomputing 2003, 55, 57–77. [Google Scholar] [CrossRef]
  35. Xu, Y.; Guo, R.; Wang, L. A Twin Multi-Class Classification Support Vector Machine. Cogn. Comput. 2013, 5, 580–588. [Google Scholar] [CrossRef]
  36. Nasiri, J.A.; Moghadam Charkari, N.; Jalili, S. Least squares twin multi-class classification support vector machine. Pattern Recogn. 2015, 48, 984–992. [Google Scholar] [CrossRef]
  37. Mangasarian, O.L. Nonlinear Programming; SIAM Press: Philadelphia, PA, USA, 1993. [Google Scholar]
  38. Tomar, D.; Agarwal, S. A comparison on multi-class classification methods based on least squares twin support vector machine. Knowl.-Based Syst. 2015, 81, 131–147. [Google Scholar] [CrossRef]
  39. Friedman, M. The use of ranks to avoid the assumption of normality implicit in the analysis of variance. J. Am. Stat. Assoc. 1937, 200, 675–701. [Google Scholar] [CrossRef]
  40. Yu, Z.; Wang, Z.; You, J.; Zhang, J.; Liu, J.; Wong, H.S.; Han, G. A New Kind of Nonparametric Test for Statistical Comparison of Multiple Classifiers Over Multiple Datasets. IEEE Trans. Cybern. 2017, 47, 4418–4431. [Google Scholar] [CrossRef]
  41. Hatamlou, A. Black hole: A new heuristic optimization approach for data clustering. Inform. Sci. 2013, 222, 175–184. [Google Scholar] [CrossRef]
  42. Chen, W.; Shao, Y.; Xu, D.; Fu, Y. Manifold proximal support vector machine for semi-supervised classification. Appl. Intell. 2014, 40, 623–638. [Google Scholar] [CrossRef]
  43. Li, Y.; Sun, H.; Yan, W.; Cui, Q. R-CTSVM+: Robust capped L1-norm twin support vector machine with privileged information. Inform. Sci. 2021, 574, 12–32. [Google Scholar] [CrossRef]
Figure 1. The flowchart for the training and predicting procedures of the proposed MNP-KSVC model.
Figure 1. The flowchart for the training and predicting procedures of the proposed MNP-KSVC model.
Information 12 00515 g001
Figure 2. The learning times on benchmark datasets for linear classifiers.
Figure 2. The learning times on benchmark datasets for linear classifiers.
Information 12 00515 g002
Figure 3. The learning times on benchmark datasets for nonlinear classifiers.
Figure 3. The learning times on benchmark datasets for nonlinear classifiers.
Information 12 00515 g003
Table 1. Statistics for benchmark datasets used in experiments. # denotes the corresponding quantity.
Table 1. Statistics for benchmark datasets used in experiments. # denotes the corresponding quantity.
Datasets#Instances#Training#Testing#Attributes#Class
Balance62543818743
Ecoli3272299875
Iris1501054543
Glass21415064136
Wine17812553133
Thyroid2151506553
Dermatology358251107346
Shuttle2175152265395
Contraceptive1473103144293
Pen Based11007703301610
Table 2. Performance comparison on benchmark datasets for linear classifiers, in terms of mean ± std of the testing multiclass accuracy (%). The Win/Loss (W/L) denotes the number of datasets for which MNP-KSVC is superior/inferior to the compared classifiers. Ave. Acc and rank denote each classifier’s average accuracy and rank over all datasets.
Table 2. Performance comparison on benchmark datasets for linear classifiers, in terms of mean ± std of the testing multiclass accuracy (%). The Win/Loss (W/L) denotes the number of datasets for which MNP-KSVC is superior/inferior to the compared classifiers. Ave. Acc and rank denote each classifier’s average accuracy and rank over all datasets.
DatasetsMulti-SVMMBSVMMTPMSVMTwin-KSVCMNP-KSVC
Balance79.91 ± 4.8686.14 ± 4.6387.43 ± 3.5686.64 ± 1.9288.33 ± 2.03
Ecoli72.32 ± 7.0373.62 ± 4.9573.77 ± 4.3374.62 ± 3.8675.91 ± 2.52
Iris93.24 ± 2.6692.96 ± 2.2493.21 ± 2.5392.69 ± 3.2494.13 ± 2.49
Glass69.18 ± 9.8572.89 ± 7.2868.68 ± 6.7971.65 ± 5.9671.38 ± 5.38
Wine93.24 ± 3.0295.28 ± 1.4698.19 ± 1.6197.23 ± 1.1297.14 ± 1.27
Thyroid90.24 ± 2.5393.74 ± 1.5892.92 ± 1.4396.97 ± 1.0897.52 ± 1.54
Dermatology81.82 ± 3.7986.67 ± 1.6984.46 ± 3.2990.37 ± 2.3289.06 ± 3.16
Shuttle71.58 ± 4.7884.04 ± 2.9277.17 ± 3.378.76 ± 4.8183.16 ± 1.92
Contraceptive38.53 ± 3.7643.95 ± 2.5144.65 ± 2.9342.25 ± 2.4544.22 ± 2.03
Pen Based79.59 ± 3.7585.94 ± 1.2681.94 ± 2.0483.21 ± 2.7786.78 ± 1.37
Ave. Acc76.9781.5280.2481.4382.76
W/L10/08/28/29/1/
Ave. rank4.62.93.12.671.7
Table 3. Performance comparison on benchmark datasets for nonlinear classifiers, in terms of the mean ± std of the testing multiclass accuracy (%). The Win/Loss (W/L) denotes the number of datasets for which MNP-KSVC is superior/inferior to the compared classifiers. Ave. Acc and rank denote each classifier’s average accuracy and rank over all datasets.
Table 3. Performance comparison on benchmark datasets for nonlinear classifiers, in terms of the mean ± std of the testing multiclass accuracy (%). The Win/Loss (W/L) denotes the number of datasets for which MNP-KSVC is superior/inferior to the compared classifiers. Ave. Acc and rank denote each classifier’s average accuracy and rank over all datasets.
DatasetsMulti-SVMMBSVMMTPMSVMTwin-KSVCMNP-KSVC
Balance79.94 ± 5.5787.13 ± 4.5789.92 ± 3.2790.17 ± 4.0891.41 ± 3.42
Ecoli79.81 ± 5.1980.82 ± 4.2584.74 ± 3.4980.12 ± 4.4984.59 ± 3.83
Iris90.26 ± 2.6596.89 ± 1.8397.39 ± 2.1494.36 ± 2.0398.04 ± 1.47
Glass58.66 ± 4.7652.64 ± 4.0862.73 ± 3.9556.01 ± 4.2664.12 ± 2.98
Wine94.36 ± 2.1294.31 ± 1.8798.38 ± 2.6297.26 ± 1.6198.04 ± 1.45
Thyroid91.82 ± 1.7593.27 ± 1.9595.34 ± 0.8994.14 ± 1.0595.63 ± 0.84
Pen Based86.53 ± 3.9389.51 ± 3.2985.06 ± 3.6886.12 ± 2.4788.78 ± 2.68
Dermatology84.43 ± 4.2983.82 ± 3.8684.51 ± 2.6483.33 ± 3.2985.26 ± 3.12
Shuttle74.36 ± 3.3983.74 ± 2.7387.06 ± 2.8986.91 ± 2.3889.37 ± 1.73
Contraceptive42.09 ± 4.9544.28 ± 4.0245.93 ± 3.8547.51 ± 3.5747.47 ± 3.68
Ave. Acc78.2280.6483.1181.5984.27
W/L10/09/18/29/1/
Ave. rank4.33.692.33.31.4
Table 4. Results of Friedman test on learning results.
Table 4. Results of Friedman test on learning results.
Statistic X F 2 p-ValueHypothesis
Linear6.51432.9503 × 10−4reject
Nonlinear10.20041.2267 × 10−5reject
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Du, S.-W.; Zhang, M.-C.; Chen, P.; Sun, H.-F.; Chen, W.-J.; Shao, Y.-H. A Multiclass Nonparallel Parametric-Margin Support Vector Machine. Information 2021, 12, 515. https://0-doi-org.brum.beds.ac.uk/10.3390/info12120515

AMA Style

Du S-W, Zhang M-C, Chen P, Sun H-F, Chen W-J, Shao Y-H. A Multiclass Nonparallel Parametric-Margin Support Vector Machine. Information. 2021; 12(12):515. https://0-doi-org.brum.beds.ac.uk/10.3390/info12120515

Chicago/Turabian Style

Du, Shu-Wang, Ming-Chuan Zhang, Pei Chen, Hui-Feng Sun, Wei-Jie Chen, and Yuan-Hai Shao. 2021. "A Multiclass Nonparallel Parametric-Margin Support Vector Machine" Information 12, no. 12: 515. https://0-doi-org.brum.beds.ac.uk/10.3390/info12120515

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop