Next Article in Journal
Phase Entropy Analysis of Electrohysterographic Data at the Third Trimester of Human Pregnancy and Active Parturition
Previous Article in Journal
Evolution Equations for Quantum Semi-Markov Dynamics
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multi-Label Feature Selection Based on High-Order Label Correlation Assumption

1
College of Computer Science and Technology, Jilin University, Changchun 130012, China
2
Key Laboratory of Symbolic Computation and Knowledge Engineering of Ministry of Education, Jilin University, Changchun 130012, China
3
College of Chemistry, Jilin University, Changchun 130012, China
*
Author to whom correspondence should be addressed.
Submission received: 14 June 2020 / Revised: 12 July 2020 / Accepted: 20 July 2020 / Published: 21 July 2020
(This article belongs to the Section Information Theory, Probability and Statistics)

Abstract

:
Multi-label data often involve features with high dimensionality and complicated label correlations, resulting in a great challenge for multi-label learning. Feature selection plays an important role in multi-label learning to address multi-label data. Exploring label correlations is crucial for multi-label feature selection. Previous information-theoretical-based methods employ the strategy of cumulative summation approximation to evaluate candidate features, which merely considers low-order label correlations. In fact, there exist high-order label correlations in label set, labels naturally cluster into several groups, similar labels intend to cluster into the same group, different labels belong to different groups. However, the strategy of cumulative summation approximation tends to select the features related to the groups containing more labels while ignoring the classification information of groups containing less labels. Therefore, many features related to similar labels are selected, which leads to poor classification performance. To this end, Max-Correlation term considering high-order label correlations is proposed. Additionally, we combine the Max-Correlation term with feature redundancy term to ensure that selected features are relevant to different label groups. Finally, a new method named Multi-label Feature Selection considering Max-Correlation (MCMFS) is proposed. Experimental results demonstrate the classification superiority of MCMFS in comparison to eight state-of-the-art multi-label feature selection methods.

1. Introduction

1.1. The Background of Multi-Label Feature Selection

During the past decade, multi-label learning has gradually attracted significant attentions and has been widely utilized in diverse real-world applications, such as text categorization [1,2], information retrieval [3,4] and gene function classification [5,6]. In multi-label data sets, each instance is related to multiple class labels simultaneously. For example, in text categorization tasks, a news document may associate with several topics simultaneously, such as “society”, “economy” and “legality”. Let X = R d denote the d-dimensional instance space and L = { l 1 , l 2 , . . , l q } denote the label space including q possible class labels. The task of multi-label learning is to obtain the set of labels related to the unseen instance x X by learning a classification model from the training data set D = { ( x 1 , L 1 ) , ( x 2 , L 2 ) , , ( x n , L n ) } , where L i L is the set of labels associated with x i and x i X ( 1 i n ) is a d-dimensional vector ( x i 1 , x i 2 , , x i d )  [7,8,9]. The classification performance of multi-label learning is closely related to the quality of input features. Like traditional single-label learning algorithm, the multi-label learning often faces with the curse of dimensionality [10].
The high-dimensional multi-label data set often contains a large number of irrelevant and redundant features that bring many disadvantages to the multi-label learning such as the computational burden and over-fitting  [10,11,12]. To address this problem, many multi-label feature selection techniques have been proposed to select the informative feature subset from the original feature set and to discard irrelevant and redundant features [13,14,15]. Feature selection techniques not only reduce the computing costs but also improve the classification performance effectively [16].
Multi-label feature selection methods are usually categorized into three groups: filter methods, wrapper methods and embedded methods [12,17,18,19,20]. Among them, filter methods are classifier-independent, that is, filter methods do not consider any learning algorithm; wrapper methods evaluate the importance of feature subsets based on the classification performance of a specific classifier; embedded methods embed the feature selection in the training process of the classifier. Filter methods have the advantage of low computational cost. In this paper, we focus on filter-based multi-label feature selection methods. In addition, filter methods rank features according to their relevance with the label set. Wrapper methods consider all possible subsets of feature combinations by using the prediction performance of a classifier to assess the quality of feature subsets. Then, the feature subset selected by wrapper methods is optimal for the learning algorithm. The disadvantage of filter methods is that its classification performance is not as good as the wrapper methods, especially in the multi-label feature selection. Therefore, we design a new method to consider the high-order label correlations and to select the most informative features for improving the prediction performance of filter methods.

1.2. Information-Theoretical-Based Multi-Label Feature Selection Methods

Different from single-label feature selection methods that evaluate the relevancy between features and only one class label (binary or multiclass), multi-label feature selection methods consider the correlations between features and a set of labels [21,22]. Moreover, the labels in multi-label data are usually not independent, where the internal correlations among labels are always very complicated [23,24]. Many filter-based feature selection methods have been proposed to take into account the label correlations on the evaluation of features, in which information-theoretical-based measures have shown to be adequate [25,26,27,28]. The purpose of the information-theoretical-based multi-label feature selection methods is to obtain an optimal feature subset by employing the information measures in information theory, where mutual information is widely utilized to evaluate the correlation between features and the label set. Suppose that S = { f 1 , f 2 , , f k } is a feature subset and L = { l 1 , l 2 , , l q } is the target label set, the mutual information I ( S ; L ) can be denoted as:
I ( S ; L ) = I ( f 1 , f 2 , , f k ; l 1 , l 2 , , l q ) = f 1 f k l 1 l q p ( f 1 , , f k , l 1 , . . l q ) log p ( f 1 , , f k , l 1 , . . l q ) p ( f 1 , , f k ) p ( l 1 , . . l q ) .
The feature subset maximizing Equation (1) provides the maximal information for the label set, which can be considered as the optimal feature subset. However, according to (1), an inevitable problem is that the joint probability p ( . ) is difficult to estimate accurately. Therefore, many feature selection methods based on low-order label correlations have emerged to obtain the approximate optimal feature subset. Some multi-label feature selection methods  [29,30,31] use the accumulated mutual information between candidate features and each label to evaluate the feature correlations, where these methods consider first-order label correlations, indicating that those labels are independent of each other. Additionally, some methods  [32,33] employ the accumulated conditional mutual information or the interaction information to measure the impact of a candidate feature with each pair of labels, considering second-order label correlations. These methods have been proved to be effective in addressing the curse of dimensionality issues. In fact, there always exist high-order label correlations in the label set that can be abstracted into several semantic groups, in which the same semantic group consists of similar labels and different semantic groups have low dependency. Thus, the cumulative summation approximation based on the whole label set may lead to the following issues:
  • Overestimating the significance of some features when these features have strong correlations with one semantic group containing many labels while being almost independent of the other labels, especially in data with a large collection of labels.
  • Ignoring the key features that are highly correlated with the semantic groups containing less labels.
  • Selecting more redundant features that are often associated with labels in the same semantic group.
In order to address the issues above, we propose a new feature selection method. The main contributions are as follows:
  • A new term named Max-Correlation (MC) is designed based on the assumption that labels cluster into several groups, the labels in the same group possess the similar semantic meaning. The MC term employs the maximum operation to select the most informative feature. Additionally, the MC term is not limited to the number of labels in the semantic group, which can effectively address the above issues numbered as 1 and 2.
  • We propose a novel feature selection method for multi-label learning based on the Max-Correlation named Multi-label Feature Selection considering the Max-Correlation (MCMFS), which not only maximizes the feature correlation between candidate features and the label set, but also minimizes the feature redundancy in the already-selected feature subset. As a result, our method intends to select the features that are from different semantic groups.
  • The effectiveness of the proposed MCMFS method is validated on one artificial data set and twelve real-world multi-label data sets. The experimental results demonstrate that the proposed method can select compact feature subsets and to achieve better classification performance in terms of multiple evaluation criteria.
The remainder of this paper is organized as follows. Section 2 introduces some basic concepts of information theory and four evaluation criteria for multi-label classification performance. Section 3 briefly reviews the related work. In Section 4, we propose the new multi-label feature selection method MCMFS. Section 5 presents the experimental results to verify the effectiveness of the proposed method. In Section 6, we draw conclusions and give the directions of our future research.

2. Preliminaries

2.1. The Basic Concepts of Information Theory

In this subsection, we introduce some basic concepts of information theory which are used to measure the correlations among random variables [34,35]. Let X = { x 1 , x 2 , , x n } and Y = { y 1 , y 2 , , y m } be two discrete random variables. The mutual information measures the amount of information shared by two variables. It is defined as follows:
I ( X ; Y ) = H ( X ) H ( X | Y ) = i = 1 n j = 1 m p ( x i , y j ) log p ( x i , y j ) p ( x i ) p ( y j ) ,
where p ( x i , y j ) is the joint probability of ( x i , y j ) , p ( x i ) is the probability of x i , p ( y j ) is the probability of y j and the base of log is 2. H ( X ) is the entropy of the variable X, which measures the uncertainty of X. H ( X | Y ) is the conditional entropy of X given Y, which measures the uncertainty left of X under the condition of Y. Mutual information can be expressed as the uncertainty reduction about variable X, given Y. H ( X ) and H ( X | Y ) are defined as:
H ( X ) = i = 1 n p ( x i ) log p ( x i ) ,
H ( X | Y ) = i = 1 n j = 1 m p ( x i , y j ) log p ( x i | y j ) ,
where p ( x i | y j ) is the conditional probability of x i given y j .
Conditional mutual information measures the mutual information between two random variables under the condition of another random variable, which is defined as:
I ( X ; Y | Z ) = H ( X | Z ) H ( X | Y , Z ) ,
where Z is a discrete random variable and H ( X | Z ) and H ( X | Y , Z ) are two conditional entropies. The joint mutual information can be defined as:
I ( X , Y ; Z ) = H ( Z ) H ( Z | X , Y ) = I ( X ; Z | Y ) + I ( Y ; Z ) .
Interaction information measures the amount of information shared by three variables, which is defined as:
I ( X ; Y ; Z ) = I ( X ; Z ) + I ( Y ; Z ) I ( X , Y ; Z ) .

2.2. Multi-Label Evaluation Metrics

To evaluate the classification performance of different multi-label feature selection methods, four evaluation metrics are widely used in multi-label learning in this paper, which are Hamming Loss, Zero-One Loss, Macro-average and Micro-average [36].
Let D = { ( x 1 , L 1 ) , ( x 2 , L 2 ) , , ( x n , L n ) } be a multi-label test set and L = { l 1 , l 2 , , l q } be the label set, where n is the number of instances and L i L is the label set corresponding to the instance x i . Suppose that L i is the predicted label set corresponding to the x i instance obtained by multi-label classifier.
Hamming Loss (HL) calculates the average fraction of misclassified labels. HL is defined as:
H L = 1 n i = 1 n | L i L i | q ,
where ⊕ denotes the symmetric difference between the label sets L i and L i . For example, let L = { l 1 , l 2 , l 3 , l 4 , l 5 } . Suppose that L i = { l 1 , l 3 , l 5 } and L i = { l 1 , l 2 , l 5 } . L i corresponds to vector v = ( 1 , 0 , 1 , 0 , 1 ) where v j = 1 or 0 ( j = 1 , 2 , . . , 5 ) means that l j is included or not included in L i . L i corresponds to vector v = ( 1 , 1 , 0 , 0 , 1 ) . Then, | L i L i | = | v v | = 2 , where ⊕ is true if v j v j .
Zero-One Loss (ZOL) calculates the average fraction of instances whose most confident label is not in the relevant label set. The definition for ZOL is:
Z O L = 1 n i = 1 n δ ( a r g m a x l L h ( x i , l ) ) ,
where δ = 1 if a r g m a x l L h ( x i , l ) L i and 0 otherwise. h ( x i , l ) is the real-valued function based on the multi-label classifier, which returns the confidence of label l being proper label of x i . a r g m a x l L h ( x i , l ) corresponds to the most confident label for x i .
Macro-average (Macro-F1) and Micro-average (Micro-F1) based on the F1 score are two widely adopted evaluation criteria for multi-label learning. Macro-F1 is an arithmetic average of the F1 score of all q labels. Macro-F1 can be obtained as follows:
M a c r o F 1 = 1 q i = 1 q 2 T P i 2 T P i + F P i + F N i ,
where T P i , F P i and F N i denote the number of true positives, false positives and false negatives in the i-th label, respectively. Micro-F1 can be considered as a weighted average of the F1 over all q labels:
M i c r o F 1 = i = 1 q 2 T P i i = 1 q ( 2 T P i + F P i + F N i ) .
The multi-label classification performance can be measured using the mentioned above evaluation criteria. For the four evaluation criteria, a lower value of HL and ZOL indicates a better classification performance. On the other hand, the higher the Macro-F1 and Micro-F1 values are, the better the classification performance is.

3. Related Work

Conventional multi-label feature selection methods can be divided into two groups to deal with the multi-label data sets: problem transformation and algorithm adaptation [37,38]. The problem transformation methods include two steps: (1) transform the multi-label data set to numerous single-label data sets; (2) select the relevant features from the transformed data sets. Binary Relevance (BR) [39], Label Power set (LP) [40] and Pruned Problem Transformation (PPT) [41] are common problem transformation methods. BR decomposes the multi-label data set into several independent binary classification data sets. LP assigns each instance’s label set to a single new class. N. Spolaôr et al. [42] propose four multi-label feature selection methods based on BR and LP which employ ReliefF (RF) [43] and Information Gain (IG) [44] as the feature evaluation criteria to measure the transformed data (RF-BR, RF-LP, IG-BR and IG-LP). However, BR ignores the label correlations and LP may create too many classes causing over-fitting and imbalance problems. PPT removes the instances with rarely occurring labels by a predefined minimal number of occurrences of the label set to improve the effectiveness of LP. Doquire and Verleysen [45] propose a multi-label feature selection method based on mutual information using PPT (PPT + MI). In addition, χ 2 statistics are used to select the effective features (PPT + CHI) [41]. However, the problem transformation-based multi-label feature selection methods usually ignore the correlations among labels or lose the label information.
In recent years, many algorithm adaptation-based multi-label feature selection methods that directly select features from the multi-label data set have been proposed. S Kashef and H Nezamabadi-pour [15] propose a multi-label feature selection algorithm based on the Pareto dominance concept that intends to select the label-specific features in multi-objective optimization problem. Sun et al. [26] propose a novel Mutual-Information-based feature selection method via constrained Convex Optimization (MICO), which obtains the discriminative features considering the label correlation. Multi-label Informed Feature Selection (MIFS) [46] is an embedded-based feature selection method that decomposes the multi-label information into a low-dimensional label space using Latent Semantic Indexing (LSI) and then employs the reduced label space to steer the feature selection process via a regression model. Lee and Kim  [32] propose a multi-label feature selection method based on information theory named Pairwise Multi-label Utility (PMU). Its evaluation function is defined as follows:
J ( f k ) = l i L I ( f k ; l i ) f j S l i L I ( f k ; f j ; l i ) l i L l j L I ( f k ; l i ; l j ) ,
where f k is a candidate feature, S is an already-selected feature subset and f j is a member of S, l i and l j are two members of the label set L. The PMU method selects the feature f k with the largest value of J ( f k ) . Multi-label feature selection method using interaction information (D2F) [29] is proposed to measure the feature correlation between features and each label in the label set. The criterion of D2F is defined as follows:
J ( f k ) = l i L I ( f k ; l i ) f j S l i L I ( f k ; f j ; l i ) .
In addition, Scalable Criterion for a Large Label Set (SCLS) [30] is proposed to design a new multi-label feature selection method based on scalable relevance evaluation. It is denoted as follows:
J ( f k ) = l i L I ( f k ; l i ) f j S I ( f k ; f j ) H ( f k ) l i L I ( f k ; l i ) .
Lin et al. [31] propose a multi-label feature selection method based on Max-Dependency and Min-Redundancy (MDMR) that maximizes the feature dependency between candidate features and each label using mutual information and minimizes the feature redundancy between candidate feature and each already-selected feature. The criterion of MDMR is denoted as follows:
J ( f k ) = l i L I ( f k ; l i ) 1 | S | f j S { ( I ( f k ; f j ) l i L I ( f k ; l i | f j ) } ,
where | S | is the number of features in the already-selected feature subset S. In addition, multi-label Feature Selection based on Label Redundancy (LRFS) [33] is proposed, and LRFS employs the conditional mutual information between candidate features and each label given other labels to consider the measurement of feature relevancy. It is defined as follows:
J ( f k ) = l i L { l i l j , l j L I ( f k ; l j | l i ) 1 | S | f j S I ( f k ; f j ) } .
Through the above introduction, we can find that previous information-theoretical-based multi-label feature selection methods employ the cumulative summation approximation to take first-order and second-order label correlations into account. In fact, there exist high-order label correlations in the real-world multi-label data sets, naturally, labels cluster into several groups. The common limitation of these methods is that the cumulative summation may overestimate the significance of some candidate features that are related to the groups containing more labels while ignoring the classification information of groups containing few labels. To explore and exploit accurately high-order correlations among labels, we first design a Max-Correlation (MC) term based on the assumption that similar labels cluster into the same groups and dissimilar labels belong to different groups. Then, we propose a novel method named Multi-label Feature Selection considering the Max-Correlation (MCMFS).

4. MCMFS: Multi-Label Feature Selection Considering the Max-Correlation

4.1. Proposed Method

Many information-theoretical-based multi-label feature selection methods apply various low-order approximations to evaluate the candidate features. D2F, SCLS and MDMR methods [29,30,31] employ the accumulated mutual information to quantify the contribution of features to the label set. The specific equation is as follows:
R e l = l i L I ( f k ; l i ) .
Equation (17) assumes that labels are independent of each other in the design of the feature relevancy term, which can be described as shown in Figure 1a, where f k is a candidate feature and l i L ( i = 1 , 2 , , q ) is one label. In addition, conditional mutual information and interaction information are also used to consider the impact of candidate features with each pair of labels ( l i , l j ) , such as PMU [32] and LRFS [33], which can be described as shown in Figure 1b.
R e l = l i L l j L I ( f k ; l i ; l j ) ,
R e l = l i L l i l j , l j L I ( f k ; l j | l i ) .
Figure 1 displays the first-order correlations and the second-order correlations among labels. However, label correlations are complicated and of high-order nature in real-world data sets. The labels can naturally cluster into several abstracted semantic meanings. For example, in text categorization, the topics “Athletics”, “Gymnastics” and “Swimming” can be extracted as the semantic meaning “Sports”, and the topics “Beach”, “Sea” and “Mountain” can be extracted as the semantic meaning “Nature”. The labels in the same semantic group have larger dependency, while labels in different semantic meanings are more distinctive. In the literature [25], the 45 labels in the medical data set, which has been used in Computational Medicine Centers 2007 Medical Natural Language Processing Challenge, are divided into 4 main groups according to the statistical information about the labels. Different groups are almost independent of each other and the number of labels in different groups is not equal. Actually, we expect to select the features that are highly discriminating to each semantic group, thereby obtaining the representative features for different semantic meaning.
Like the Equations (17)–(19), the cumulative summation of information terms tends to select features that are related to one semantic group, which leads to overestimating some feature significance especially when the number of labels in the same semantic group is large. As a result, many redundant features are selected. For example, suppose that the total number of labels is 100 and there are two main semantic groups in the label set, that are C 1 and C 2 . If the number of labels in C 1 is 90 and the number of labels in C 2 is 10, then the cumulative summation criterion prefers to select the features that are associated with the labels in the semantic group C 1 , while reducing the selection possibility of features that are from C 2 . In such a situation, the critical features are neglected that are highly related to the semantic groups containing few labels because the value of the cumulative summation will be small when these features are independent of most other labels. Additionally, the selection possibility of redundant features will increase due to the overestimation of the features significance when these features are associated with the same semantic group that contains many labels. However, an effective and compact feature subset should choose features that are from different semantic groups, which is proved to be effective [47].
To address the issue, we propose a new multi-label feature selection method to select features that are from different semantic groups. Suppose that the label set L = { l 1 , l 2 , , l q } can be divided into m semantic groups, that is, L = { C 1 , C 2 , , C m } , where each semantic group C i = { l i 1 , l i 2 , , l i q i } L ( i = 1 , 2 , , m ) and it satisfies C 1 C 2 C m = L and C i C j = . Our aim is to select the critical features that are from each semantic group, which is described as shown in Figure 2.
In order to avoid the overestimation problem caused by the number of labels in the different semantic groups, we employ the maximum operation to measure the mutual information between the candidate feature f k and each semantic group C i ( i = 1 , 2 , . . , m ) . The specific equation is as follows:
R ( f k , C i ) = max l j C i ( I ( f k ; l j ) ) .
Equation (20) measures the relevancy between the candidate feature and labels in the same semantic group. The larger the value of Equation (20), the more important the candidate feature is in the semantic group. Equation (20) is the upper bound of the maximal relevancy between one candidate feature and the labels in the semantic group. Furthermore, a small value of Equation (20) means that the relevancy between the candidate feature and the labels in the semantic group is weak. Finally, Equation (20) can effectively avoid the overestimation caused by accumulation, even if many labels are in the same semantic group.
Thereafter, according to Equation (20), an m-dimensional vector Cor ( f k ; L ) of feature f k and the label set L is obtained, that is Cor ( f k ; L ) = [ R ( f k , C 1 ) , R ( f k , C 2 ) , , R ( f k , C m ) ] . We select the maximum value of Cor ( f k ; L ) as the feature relevancy between candidate features and the entire label set L, which is named Max-Correlation (MC). It is defined as:
M C ( f k ; L ) = max C i L R ( f k , C i ) = max C i L { max l j C i I ( f k ; l j ) } = max l j L I ( f k ; l j ) .
M C ( f k ; L ) can effectively capture the maximum amount of contribution of the feature regarding the label set. Meanwhile, M C ( f k ; L ) can accurately select the critical features whatever how many labels in the semantic groups. Based on the definition of Max-Correlation, we propose a novel multi-label feature selection method named Multi-label Feature Selection considering the Max-Correlation (MCMFS). The evaluation function is as follows:
J ( f k ) = M C ( f k ; L ) 1 | S | f j S I ( f k ; f j ) = max l j L I ( f k ; l j ) 1 | S | f j S I ( f k ; f j ) .
I ( f k ; f j ) measures the feature redundancy between the candidate feature f k and each already-selected feature f j . 1 | S | is employed to balance the magnitude between the Max-Correlation term and the feature redundancy term. Therefore, Equation (22) uses M C ( f k ; L ) to maximize the feature relevancy between candidate features and the label set, while using the mutual information I ( f k ; f j ) to minimize the feature redundancy in the already-selected feature subset to choose the feature that are from different semantic groups. The sequential forward search strategy is used in the process of feature selection. We select the feature f k that achieves the maximum value of J ( f k ) as the next already-selected feature. The pseudo code of MCMFS is presented in Algorithm 1.
Algorithm 1 MCMFS
Input:
  A training sample D with a full feature set F = { f 1 , f 2 , , f d } and the label set L = { l 1 , l 2 , , l q } ;
  The number of selected features b.
Output:
  The already-selected feature subset S.
 1:
S ;
 2:
a 0 ;
 3:
for  i = 1 to d do
 4:
 According to the Equation (21) calculate the M C ( f i ; L ) ;
 5:
end for
 6:
while  a < b   do
 7:
if a == 0  then
 8:
  Select the feature f j with the largest M C ( f i ; Y ) ;
 9:
   a = a + 1 ;
10:
   S = S { f j } ;
11:
   F = F { f j } ;
12:
end if
13:
for each candidate feature f i F do
14:
  Calculate the mutual information I ( f i ; f j ) ;
15:
  According to the Equation (22) calculate the J ( f i ) ;
16:
end for
17:
 Select the feature f j with the largest J ( f i ) ;
18:
S = S { f j } ;
19:
F = F { f j } ;
20:
a = a + 1 ;
21:
end while
There are three stages in the MCMFS method. In the first stage (lines 1–5), it initializes the parameters, which includes the already-selected feature subset S and the number of already-selected features a in lines 1–2, and calculates the Max-Correlation for each feature in lines 3–5. The second stage (lines 7–12) selects the maximum value of Max-Correlation as the first already-selected feature. The third stage (lines 13–20) calculates the Equation (22) and selects the next feature.
The minimal-redundancy-maximum-relevance (mRMR) [48] is a well-known single-label feature selection method, which uses mutual information between candidate features and class labels to evaluate feature relevance and adopts the same feature redundancy term with our method. The resemblance between mRMR method and MCMCFS method is that both methods consider the relationship between candidate features and already-selected features to minimize feature redundancy. The difference is that mRMR method does not consider the effects of label correlations. In multi-label feature selection, the proposed MCMCFS method employs the Max-Correlation term to consider the high-order label correlations.

4.2. Complexity Analysis

We provide the complexity analysis for the MCMFS method and other five information-theoretical-based feature selection methods (D2F, PMU, SCLS, MDMR and LRFS). Suppose that the number of instances is n, the number of features is d and the number of labels is q. The mutual information, conditional mutual information and interaction information need the time complexity of O ( n ) since all the instances need to be visited for probability estimation. Suppose that the number of selected features is b, then the time complexity of MCMFS and SCLS is O ( n d q + b n d ) . The time complexity of D2F and MDMR is O ( n d q + b n d q ) . PMU and LRFS design the evaluation criteria to consider the second-order label correlations. The time complexity of PMU is O ( n d q + b n d q + n d q 2 ) and the time complexity of LRFS is O ( n d q 2 + b n d ) . Table 1 lists the time complexity of these methods. As shown in Table 1, MCMFS achieves the same time complexity with SCLS method. In addition, the time complexity of MCMFS method is lower than that of the D2F, MDMR, PMU and LRFS methods. Therefore, the proposed method is more computationally efficient than these four methods.

5. Experimental Results and Analysis

In this section, we evaluate the classification performance of the proposed MCMFS method and present the experimental results. MCMFS is compared to one embedded-based method (MIFS [46]) and two problem transformation-based methods (PPT + MI [45] and PPT + CHI [41]) and five information-theoretical-based methods (D2F [29], MDMR [31], PMU [32], SCLS [30] and LRFS [33]). First, we introduce the experimental settings and describe the evaluation framework in Figure 3. Second, MCMFS is compared to five information-theoretical-based methods that employ the cumulative summation approximation to evaluate the candidate features on an artificial data set. Finally, the MCMFS method is compared to the eight representative methods on 12 real-word multi-label data sets in terms of four evaluation metrics to verify the effectiveness of MCMFS method. All the experiments are executed on an Intel Core (TM) i7-6700 with 3.4 GHz processing speed.

5.1. Experimental Setting

The experimental setting is as follows: First, the continuous features are discretized using an equal-width strategy into three bins, as recommend in the literature [14,29]. Second, the number of already-selected features b varies from 1 to M with a step size of 1, where M is 20% of the total number of features (M = 17% in medical data set used in experiments). Third, we employ the MLKNN [47] as the multi-label classifier to evaluate the classification performance of the MCMFS method and other eight compared feature selection methods in terms of Hamming Loss and Zero-One Loss. Additionally, the number of nearest neighbors K is set to 10. Finally, k-Nearest Neighbors (kNN) and Liblinear-based Support Vector Machine (SVM) are implemented to evaluate the classification performance in terms of Macro-F1 and Micro-F1. The kNN is a non-linear neighborhood-based classifier, while the SVM is a linear classifier. We adopt two different classifiers to display the different classification performance of these methods. In addition, kNN and SVM are two popular classifiers in the feature selection methods based on information theory, they are widely applied in various literature [49,50,51,52,53]. Different k values of kNN classifier appear to have less effect on the classification performance for the filter methods [53]. In these references, k is set to 3, indicating that this is an empirical setting. Therefore, we set k to 3 in this paper. We use the package scikit-learn and in Python 2.7 to implement the classifiers. The multi-label data sets used in the experiment are from Mulan Library [46] where the training set and test set have been already separated in the data source. Therefore, as shown in Figure 3, we use the result of feature selection on the training set to implement on test set directly.

5.2. Experiment and Analysis on an Artificial Data Set

We apply an artificial data to visually compare MCMFS to five information-theoretical-based methods (D2F, LRFS, MDMR, PMU and SCLS) that employ the cumulative summation approximation to evaluate the importance of candidate features. Table 2 displays the artificial data D = { O , F , L } , where O = { o 1 , o 2 , , o 10 } , F = { f 1 , f 2 , , f 8 } and L = { l 1 , l 2 , , l 5 } .
The matrix of mutual information between labels D ( L , L ) is listed. It can be observed that l 1 , l 2 and l 3 have close correlations, and they have weak correlation with l 4 and l 5 . In addition, l 4 and l 5 are also weakly correlated with each other. For example, for the label l 1 , the values of I ( l 1 ; l 2 ) and I ( l 1 ; l 3 ) are significantly greater than the values of I ( l 1 ; l 4 ) and I ( l 1 ; l 5 ) . For the label l 4 , all the values of I ( l 4 ; l 1 ) , I ( l 4 ; l 2 ) , I ( l 4 ; l 3 ) and I ( l 4 ; l 5 ) are very small. Intuitively, the label set can cluster into three semantic groups, that are C 1 = { l 1 , l 2 , l 3 } , C 2 = { l 4 } and C 3 = { l 5 } .
D ( L , L ) = I ( l 1 ; l 1 ) I ( l 1 ; l 2 ) I ( l 1 ; l 3 ) I ( l 1 ; l 4 ) I ( l 1 ; l 5 ) I ( l 2 ; l 1 ) I ( l 2 ; l 2 ) I ( l 2 ; l 3 ) I ( l 2 ; l 4 ) I ( l 2 ; l 5 ) I ( l 3 ; l 1 ) I ( l 3 ; l 2 ) I ( l 3 ; l 3 ) I ( l 3 ; l 4 ) I ( l 3 ; l 5 ) I ( l 4 ; l 1 ) I ( l 4 ; l 2 ) I ( l 4 ; l 3 ) I ( l 4 ; l 4 ) I ( l 4 ; l 5 ) I ( l 5 ; l 1 ) I ( l 5 ; l 2 ) I ( l 5 ; l 3 ) I ( l 5 ; l 4 ) I ( l 5 ; l 5 ) = 0.971 0.256 0.256 0.006 0.006 0.256 0.971 0.256 0.006 0.006 0.256 0.256 0.971 0.006 0.006 0.006 0.006 0.006 0.881 0.002 0.006 0.006 0.006 0.002 0.881
We present the feature ranking results and the classification performance obtained by the six feature selection methods in Table 3. Five-fold cross-validation is employed to evaluate the classification performance on the artificial data. The values in bold font represent the best classification performance in Table 3. It can be seen that MCMFS obtains better experimental results in terms of HL, ZOL, Macro-F1 and Micro-F1. According to the results of feature ranking, the five compared methods give lesser importance rankings for certain features. For example, compared to D2F, LRFS, MDMR and SCLS, the rank of f 8 is higher in MCMFS. In fact, f 8 is the most relevant feature to the label l 5 ( f 8 = a r g m a x f i F ( I ( f i ; l 5 ) ) ). Compared to D2F, LRFS and PMU, the rank of f 3 is higher in MCMFS, where f 3 is the most relevant feature to the label l 4 ( f 3 = a r g m a x f i F ( I ( f i ; l 4 ) ) ). In other words, f 8 and f 3 are critical features of the semantic groups C 3 and C 2 , respectively. f 2 is the key feature of the semantic group C 1 that is selected by most methods. The proposed method finds accurately the key features that belong to different semantic groups.

5.3. Experimental Results on the Real-Word Data Sets

The experiments are conducted on 12 real-world multi-label data sets that are from Mulan Library [54]. The description of the data sets is presented in Table 4. These data sets contain different number of instances, features and labels. In addition, these data sets cover two different application areas. The data set scene is collected for semantic image categorization and the remaining data sets are widely applied to text categorization.
Table 5 and Table 6 record the average classification results and standard deviations of the proposed method and other eight compared methods on the 12 data sets in terms of Hamming Loss and Zero-One Loss, respectively. The values in bold font represent the best classification performance achieved by the corresponding method.
In Table 5, MCMFS obtains the best Hamming Loss performance on 11 data sets. MIFS method provides better results for the Business data set, which means that the decomposition process of the label set of the MIFS method is helpful for the feature selection on the Business data set. As shown in Table 6, PPT + CHI obtains better performance than the proposed MCMFS method and other compared methods on Reference data set in terms of Zero-One Loss performance. MCMFS obtains the best Zero-One Loss performance on 11 data sets. On the whole, MCMFS provides better classification performance compared to other competitive feature selection methods in terms of Hamming Loss and Zero-One Loss on MLKNN classifier.
Table 7, Table 8, Table 9 and Table 10 record the classification performance of the proposed method and other eight compared methods in terms of Macro-F1 and Micro-F1, respectively. Table 7 and Table 8 present the Macro-F1 metric on the SVM classifier and 3NN classifier, respectively. As the results indicate, we can observe that D2F obtains the best Macro-F1 performance on the enron data set using the SVM classifier in Table 7. Our method outperforms the compared methods in terms of the Macro-F1 performance on 11 data sets using the SVM classifier and on 12 data sets using the 3NN classifier. Table 9 and Table 10 show the Micro-F1 performance on the SVM classifier and 3NN classifier, respectively. Compared to the eight methods, MCMFS obtains the best Micro-F1 performance on 11 data sets using the SVM classifier. In Table 10, MCMFS obtains the best Micro-F1 performance on 9 data sets using the 3NN classifier. Overall, our method achieves the best classification performance in terms of Macro-F1 and Micro-F1 on these data sets using the SVM classifier and 3NN classifier.
Observing these results, PPT+CHI provides better classification performance on Reference data set in terms of ZOL, Macro-F1 on SVM classifier and 3NN classifier, Micro-F1 on SVM classifier. χ 2 is effective in evaluating the features of Reference data set by transforming the label set to single label using PPT. Compared with the information-theoretical-based methods, the classification performance of MCMFS is is the best among all methods, followed by LRFS, MDMR, D2F, PMU and SCLS, which verifies the effectiveness of using the maximum operation instead of the cumulative summation approximation to take into account the higher-order label relationship.
To clearly show the classification performances of different feature selection methods, Figure 4, Figure 5 and Figure 6 show the experimental results on three data sets (Arts, medical and scene). In these figures, the X-axis represents the number of already-selected features, which is varied as {1%, 2%, …, 20%} or {1%, 2%, …, 17%} (medical data set) of the total number of features. The Y-axis represents the experimental results of the different evaluation criteria. Different colors and shapes indicate different multi-label feature selection methods.
According to the classification performance on the Figure 4, Figure 5 and Figure 6, we can observe that MCMFS obtains better classification performance than other compared feature selection methods. Compared to five information-theoretical-based methods D2F, MDMR, PMU, SCLS and LRFS, the experimental results demonstrate that the maximum operation is more effective than the cumulative summation approximation operation. In addition, MCMFS outperforms the other three multi-label feature selection methods PPT + MI, PPT + CHI and MIFS on these data sets.
Finally, we show the running time of MCMFS and other eight compared methods in Table 11. The running time of PPT + MI and PPT + CHI methods is the minimum, because they only need one iteration on the transformed single label to complete the feature selection. Although SCLS and MIFS methods have lower running time than our method, the proposed method outperforms these two methods in terms of multiple evaluation criteria for the classification performance. As compared to D2F, MDMR, PMU and LRFS, our method is more computationally efficient. Therefore, the running time of MCMFS method is generally acceptable. Additionally, we use Figure 7 to present the minimum and maximum values of each method on different data sets. In Figure 7, the X-axis represents data sets while the Y-axis represents the running time of each method. To clearly show the running time of different methods, we use Figure 7b to display the running time of MCMFS, PPT + MI, PPT + CHI, MIFS and SCLS methods. As shown in Figure 7, we can find that PMU obtains the most running time among all methods. PPT + MI has the least running time. The running time of our method MCMFS is acceptable. In addition, the running time of all methods increases as the size of the data sets increases.

6. Conclusions

In this paper, a novel multi-label feature selection method is proposed named Multi-label Feature Selection considering the Max-Correlation (MCMFS). The Max-Correlation (MC) term is designed based on the high-order label correlations and the assumption that labels naturally cluster into several groups. The combination of maximum operation and the feature redundancy term contributes to selecting the features that are from different label groups.
To demonstrate the effectiveness of our method, MCMFS is compared to five information-theoretical-based multi-label feature selection methods (D2F, MDMR, PMU, SCLS and LRFS) that employ the cumulative summation approximation operation to select features on an artificial data set. Furthermore, MCMFS is compared to eight state-of-the art multi-label feature selection methods (PPT + MI, PPT + CHI, MIFS, D2F, MDMR, PMU, SCLS and LRFS) using MLKNN on 12 real-world multi-label data sets in terms of Hamming Loss and Zero One Loss. Additionally, the 3NN classifier and SVM classifier are used to evaluate the classification performance among the nine feature selection methods in terms of Macro-F1 and Micro-F1. The experimental results demonstrate that MCMFS obtains better classification results than the compared methods and can effectively select a compact feature subset for the classification.
Finally, in our future work, we intend to explore high-order label correlations and sparse learning for multi-label feature selection. Additionally, we intend to propose a method that can automatically assign the appropriate number of feature subsets to each data set.

Author Contributions

Conceptualization, P.Z. and W.G.; methodology, P.Z.; software, P.Z.; validation, J.H., Y.L.; formal analysis, P.Z.; investigation, P.Z.; resources, J.H.; data curation, Y.L.; writing—original draft preparation, P.Z.; writing—review and editing, W.G.; visualization, P.Z.; supervision, P.Z.; project administration, W.G.; funding acquisition, W.G. All authors have read and approved the final manuscript.

Funding

This research was funded by Postdoctoral Innovative Talents Support Program under Grant No. BX20190137, and China Postdoctoral Science Foundation funded project under Grant No. 2020M670839, and National Key R&D Plan of China under Grant No. 2017YFA0604500, and by National Sci-Tech Support Plan of China under Grant No. 2014BAH02F00, and by National Natural Science Foundation of China under Grant No. 61701190, and by Youth Science Foundation of Jilin Province of China under Grant No. 20160520011JH & 20180520021JH, and by Youth Sci-Tech Innovation Leader and Team Project of Jilin Province of China under Grant No. 20170519017JH, and by Key Technology Innovation Cooperation Project of Government and University for the whole Industry Demonstration under Grant No. SXGJSF2017-4, and by Key scientific and technological R&D Plan of Jilin Province of China under Grant No. 20180201103GX, Project of Jilin Province Development and Reform Commission No. 2019FGWTZC001.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Ueda, N.; Saito, K. Parametric Mixture Models for Multi-Labeled Text. In Proceedings of the International Conference on Neural Information Processing Systems, Vancouver, CO, Canada, 9–14 December 2002. [Google Scholar]
  2. Kazawa, H.; Izumitani, T.; Taira, H.; Maeda, E. Maximal margin labeling for multi-topic text categorization. In Proceedings of the Advances in Neural Information Processing Systems, Vancouver, BC, Canada, 13–18 December 2004; pp. 649–656. [Google Scholar]
  3. Gopal, S.; Yang, Y. Multilabel classification with meta-level features. In Proceedings of the 33rd International ACM SIGIR Conference on Research and Development in Information Retrieval, Geneva, Switzerland, 19 July 2010; pp. 315–322. [Google Scholar]
  4. Huang, S.J.; Zhou, Z.H. Multi-label learning by exploiting label correlations locally. In Proceedings of the Twenty-Sixth AAAI Conference on Artificial Intelligence, Toronto, ON, Canada, 22–26 July 2012. [Google Scholar]
  5. Zhang, M.L.; Zhou, Z.H. Multilabel neural networks with applications to functional genomics and text categorization. IEEE Trans. Knowl. Data Eng. 2006, 18, 1338–1351. [Google Scholar] [CrossRef] [Green Version]
  6. Elisseeff, A.; Weston, J. A kernel method for multi-labelled classification. In Proceedings of the International Conference on Neural Information Processing Systems: Natural and Synthetic, Cambridge, MA, USA, 3–8 December 2001; pp. 681–687. [Google Scholar]
  7. Zhang, M.L.; Zhou, Z.H. A Review on Multi-Label Learning Algorithms. IEEE Trans. Knowl. Data Eng. 2014, 26, 1819–1837. [Google Scholar] [CrossRef]
  8. Huang, J.; Li, G.; Huang, Q.; Wu, X. Learning label specific features for multi-label classification. In Proceedings of the 2015 IEEE International Conference on Data Mining, Atlantic City, NJ, USA, 14–17 November 2015; pp. 181–190. [Google Scholar]
  9. Huang, J.; Li, G.; Huang, Q.; Wu, X. Learning label-specific features and class-dependent labels for multi-label classification. IEEE Trans. Knowl. Data Eng. 2016, 28, 3309–3323. [Google Scholar] [CrossRef]
  10. Monard, M.C.; Tsoumakas, G.; Lee, H.D. A systematic review of multi-label feature selection and a new method based on label construction. Neurocomputing 2016, 180, 3–15. [Google Scholar]
  11. Gao, W.; Hu, L.; Zhang, P. Class-specific mutual information variation for feature selection. Pattern Recognit. 2018, 79, 328–339. [Google Scholar] [CrossRef]
  12. Lin, Y.; Hu, Q.; Liu, J.; Chen, J.; Duan, J. Multi-label feature selection based on neighborhood mutual information. Appl. Soft Comput. 2016, 38, 244–256. [Google Scholar] [CrossRef]
  13. Wang, J.J.Y.; Bensmail, H.; Gao, X. Feature selection and multi-kernel learning for sparse representation on a manifold. Neural Netw. 2014, 51, 9–16. [Google Scholar] [CrossRef]
  14. Lee, J.; Kim, D.W. Memetic feature selection algorithm for multi-label classification. Inf. Sci. 2015, 293, 80–96. [Google Scholar] [CrossRef]
  15. Kashef, S.; Nezamabadi-pour, H. A label-specific multi-label feature selection algorithm based on the Pareto dominance concept. Pattern Recognit. 2019, 88, 654–667. [Google Scholar] [CrossRef]
  16. Lin, Y.; Hu, Q.; Zhang, J.; Wu, X. Multi-label feature selection with streaming labels. Inf. Sci. 2016, 372, 256–275. [Google Scholar] [CrossRef]
  17. Lee, J.; Kim, D.W. Fast multi-label feature selection based on information-theoretic feature ranking. Pattern Recognit. 2015, 48, 2761–2771. [Google Scholar] [CrossRef]
  18. Yu, K.; Yu, S.; Tresp, V. Multi-label informed latent semantic indexing. In Proceedings of the 28th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval, Salvador, Brazil, 15–19 August 2005; pp. 258–265. [Google Scholar]
  19. Zhu, R.; Dornaika, F.; Ruichek, Y. Learning a discriminant graph-based embedding with feature selection for image categorization. Neural Netw. 2019, 111, 35–46. [Google Scholar] [CrossRef] [PubMed]
  20. Zhang, M.L.; Peña, J.M.; Robles, V. Feature selection for multi-label naive Bayes classification. Inf. Sci. 2009, 179, 3218–3229. [Google Scholar] [CrossRef]
  21. Kong, X.; Yu, P.S. gMLC: A multi-label feature selection framework for graph classification. Knowl. Inf. Syst. 2012, 31, 281–305. [Google Scholar] [CrossRef]
  22. Zhang, Y.; Zhou, Z.H. Multilabel dimensionality reduction via dependence maximization. In Proceedings of the National Conference on Artificial Intelligence, Chicago, IL, USA, 13–17 July 2008. [Google Scholar]
  23. Gu, Q.; Li, Z.; Han, J. Correlated multi-label feature selection. In Proceedings of the 20th ACM International Conference on Information and Knowledge Management, Glasgow, Scotland, UK, 24–28 October 2011; pp. 1087–1096. [Google Scholar]
  24. Suping, X.U.; Yang, X.; Yunsong, Q.I. Multi-label learning with label-specific feature reduction. J. Comput. Appl. 2016, 104, 52–61. [Google Scholar]
  25. Li, F.; Miao, D.; Pedrycz, W. Granular multi-label feature selection based on mutual information. Pattern Recognit. 2017, 67, 410–423. [Google Scholar] [CrossRef]
  26. Sun, Z.; Zhang, J.; Dai, L.; Li, C.; Zhou, C.; Xin, J.; Li, S. Mutual information based multi-label feature selection via constrained convex optimization. Neurocomputing 2019, 329, 447–456. [Google Scholar] [CrossRef]
  27. Li, L.; Liu, H.; Ma, Z.; Mo, Y.; Duan, Z.; Zhou, J.; Zhao, J. Multi-label feature selection via information gain. In Proceedings of the International Conference on Advanced Data Mining and Applications, Guilin, China, 19–21 December 2014; pp. 345–355. [Google Scholar]
  28. Sechidis, K.; Spyromitros-Xioufis, E.; Vlahavas, I. Information Theoretic Multi-Target Feature Selection via Output Space Quantization. Entropy 2019, 21, 855. [Google Scholar] [CrossRef] [Green Version]
  29. Lee, J.; Kim, D.W. Mutual Information-based multi-label feature selection using interaction information. Expert Syst. Appl. 2015, 42, 2013–2025. [Google Scholar] [CrossRef]
  30. Lee, J.; Kim, D.W. SCLS: Multi-label feature selection based on scalable criterion for large label set. Pattern Recognit. 2017, 66, 342–352. [Google Scholar] [CrossRef]
  31. Lin, Y.; Hu, Q.; Liu, J.; Duan, J. Multi-label feature selection based on max-dependency and min-redundancy. Neurocomputing 2015, 168, 92–103. [Google Scholar] [CrossRef]
  32. Lee, J.; Kim, D.W. Feature selection for multi-label classification using multivariate mutual information. Pattern Recognit. Lett. 2013, 34, 349–357. [Google Scholar] [CrossRef]
  33. Zhang, P.; Liu, G.; Gao, W. Distinguishing two types of labels for multi-label feature selection. Pattern Recognit. 2019, 95, 72–82. [Google Scholar] [CrossRef]
  34. Shannon, C.E. A mathematical theory of communication. Bell Syst. Tech. J. 1948, 27, 379–423. [Google Scholar] [CrossRef] [Green Version]
  35. Cover, T.M.; Thomas, J.A. Elements of Information Theory; Tsinghua University Press: Beijing, China, 2003; pp. 13–55. [Google Scholar]
  36. Wu, X.Z.; Zhou, Z.H. A unified view of multi-label performance measures. In Proceedings of the 34th International Conference on Machine Learning-Volume 70, Sydney, Australia, 6–11 August 2017; pp. 3780–3788. [Google Scholar]
  37. Lee, J.; Kim, D.W. Efficient multi-label feature selection using entropy-based label selection. Entropy 2016, 18, 405. [Google Scholar] [CrossRef] [Green Version]
  38. Lim, H.; Lee, J.; Kim, D.W. Optimization approach for feature selection in multi-label classification. Pattern Recognit. Lett. 2017, 89, 25–30. [Google Scholar] [CrossRef]
  39. Boutell, M.R.; Luo, J.; Shen, X.; Brown, C.M. Learning multi-label scene classification. Pattern Recognit. 2004, 37, 1757–1771. [Google Scholar] [CrossRef] [Green Version]
  40. Trohidis, K.; Tsoumakas, G.; Kalliris, G.; Vlahavas, I. Multilabel classification of music into emotions. Blood 2008, 90, 3438–3443. [Google Scholar]
  41. Read, J. A pruned problem transformation method for multi-label classification. In Proceedings of the 2008 New Zealand Computer Science Research Student Conference, Christchurch, New Zealand, 14–18 April 2008; pp. 143–150. [Google Scholar]
  42. SpolaôR, N.; Cherman, E.A.; Monard, M.C.; Lee, H.D. A comparison of multi-label feature selection methods using the problem transformation approach. Electron. Notes Theor. Comput. Sci. 2013, 292, 135–151. [Google Scholar] [CrossRef] [Green Version]
  43. Robnik-Ikonja, M.; Kononenko, I. Theoretical and Empirical Analysis of ReliefF and RReliefF. Mach. Learn. 2003, 53, 23–69. [Google Scholar] [CrossRef] [Green Version]
  44. Lewis, D.D. Feature Selection and Feature Extraction for Text Categorization. In Proceedings of the Speech and Natural Language: Proceedings of a Workshop Held at Harriman, New York, NY, USA, 23–26 February 1992; pp. 212–217. [Google Scholar]
  45. Doquire, G.; Verleysen, M. Mutual information-based feature selection for multilabel classification. Neurocomputing 2013, 122, 148–155. [Google Scholar] [CrossRef]
  46. Jian, L.; Li, J.; Shu, K.; Liu, H. Multi-Label Informed Feature Selection. In Proceedings of the Twenty-Fifth International Joint Conference on Artificial Intelligence, New York, NY, USA, 9–15 July 2016; pp. 1627–1633. [Google Scholar]
  47. Kong, D.; Fujimaki, R.; Liu, J.; Nie, F.; Ding, C. Exclusive Feature Learning on Arbitrary Structures via 1,2-norm. In Proceedings of the Advances in Neural Information Processing Systems, Montreal, QC, Canada, 8–13 December 2014; pp. 1655–1663. [Google Scholar]
  48. Peng, H.; Long, F.; Ding, C. Feature selection based on mutual information criteria of max-dependency, max-relevance, and min-redundancy. IEEE Trans. Pattern Anal. Mach. Intell. 2005, 27, 1226–1238. [Google Scholar] [CrossRef] [PubMed]
  49. Vinh, N.X.; Zhou, S.; Chan, J.; Bailey, J. Can high-order dependencies improve mutual information based feature selection? Pattern Recognit. 2016, 53, 46–58. [Google Scholar] [CrossRef] [Green Version]
  50. Bennasar, M.; Hicks, Y.; Setchi, R. Feature selection using Joint Mutual Information Maximisation. Expert Syst. Appl. 2015, 42, 8520–8532. [Google Scholar] [CrossRef] [Green Version]
  51. Hoque, N.; Bhattacharyya, D.K.; Kalita, J. MIFS-ND: A mutual information-based feature selection method. Expert Syst. Appl. 2014, 41, 6371–6385. [Google Scholar] [CrossRef]
  52. Che, J.; Yang, Y.; Li, L.; Bai, X.; Zhang, S.; Deng, C. Maximum relevance minimum common redundancy feature selection for nonlinear data. Inf. Sci. 2017, 409, 68–86. [Google Scholar] [CrossRef]
  53. Freeman, C.; Kulic, D.; Basir, O. An evaluation of classifier-specific filter measure performance for feature selection. Pattern Recognit. 2015, 48, 1812–1826. [Google Scholar] [CrossRef]
  54. Tsoumakas, G.; Spyromitros-Xioufis, E.; Vilcek, J.; Vlahavas, I. MULAN: A Java library for multi-label learning. J. Mach. Learn. Res. 2011, 12, 2411–2414. [Google Scholar]
Figure 1. The correlation between feature f k and the label set for the first-order and second-order label correlations.
Figure 1. The correlation between feature f k and the label set for the first-order and second-order label correlations.
Entropy 22 00797 g001
Figure 2. The correlation between feature f k and the label set for the high-order label correlations.
Figure 2. The correlation between feature f k and the label set for the high-order label correlations.
Entropy 22 00797 g002
Figure 3. Experimental framework.
Figure 3. Experimental framework.
Entropy 22 00797 g003
Figure 4. Classification performance on Arts data set: (a) Hamming Loss, (b) Zero-One Loss, (c) Macro-F1 on SVM classifier, (d) Macro-F1 on 3NN classifier, (e) Micro-F1 on SVM classifier, (f) Micro-F1 on 3NN classifier.
Figure 4. Classification performance on Arts data set: (a) Hamming Loss, (b) Zero-One Loss, (c) Macro-F1 on SVM classifier, (d) Macro-F1 on 3NN classifier, (e) Micro-F1 on SVM classifier, (f) Micro-F1 on 3NN classifier.
Entropy 22 00797 g004
Figure 5. Classification performance on medical data set: (a) Hamming Loss, (b) Zero-One Loss, (c) Macro-F1 on SVM classifier, (d) Macro-F1 on 3NN classifier, (e) Micro-F1 on SVM classifier, (f) Micro-F1 on 3NN classifier.
Figure 5. Classification performance on medical data set: (a) Hamming Loss, (b) Zero-One Loss, (c) Macro-F1 on SVM classifier, (d) Macro-F1 on 3NN classifier, (e) Micro-F1 on SVM classifier, (f) Micro-F1 on 3NN classifier.
Entropy 22 00797 g005
Figure 6. Classification performance on scene data set: (a) Hamming Loss, (b) Zero-One Loss, (c) Macro-F1 on SVM classifier, (d) Macro-F1 on 3NN classifier, (e) Micro-F1 on SVM classifier, (f) Micro-F1 on 3NN classifier.
Figure 6. Classification performance on scene data set: (a) Hamming Loss, (b) Zero-One Loss, (c) Macro-F1 on SVM classifier, (d) Macro-F1 on 3NN classifier, (e) Micro-F1 on SVM classifier, (f) Micro-F1 on 3NN classifier.
Entropy 22 00797 g006
Figure 7. The running time of nine methods: (a) all methods, (b) Multi-label Feature Selection considering the Max-Correlation (MCMFS), Pruned Problem Transformation (PPT) + mutual information (MI), PPT + CHI, Multi-label Informed Feature Selection (MIFS) and Scalable Criterion for a Large Label Set (SCLS).
Figure 7. The running time of nine methods: (a) all methods, (b) Multi-label Feature Selection considering the Max-Correlation (MCMFS), Pruned Problem Transformation (PPT) + mutual information (MI), PPT + CHI, Multi-label Informed Feature Selection (MIFS) and Scalable Criterion for a Large Label Set (SCLS).
Entropy 22 00797 g007
Table 1. The time complexity of six methods.
Table 1. The time complexity of six methods.
MethodsTime Complexity
MCMFS O ( n d q + b n d )
SCLS O ( n d q + b n d )
D2F O ( n d q + b n d q )
MDMR O ( n d q + b n d q )
PMU O ( n d q + b n d q + n d q 2 )
LRFS O ( n d q 2 + b n d )
Table 2. An artificial data.
Table 2. An artificial data.
O f 1 f 2 f 3 f 4 f 5 f 6 f 7 f 8 l 1 l 2 l 3 l 4 l 5
o 1 0100010011100
o 2 1000011100001
o 3 1011011000010
o 4 1100001111111
o 5 0010101000010
o 6 0100000011100
o 7 0000001000001
o 8 0001000010000
o 9 1000010001000
o 10 0000101000100
Table 3. Experimental results on the artificial data set.
Table 3. Experimental results on the artificial data set.
MethodsFeature RankingHL↓ZOL↓Macro-F1
(SVM) ↑
Micro-F1
(SVM) ↑
Macro-F1
(3NN) ↑
Micro-F1
(3NN) ↑
MCMFS { f 2 , f 8 , f 3 , f 7 , f 4 , f 5 , f 6 , f 1 } 0.30750.88750.40170.46510.37830.4694
D2F { f 2 , f 7 , f 1 , f 6 , f 8 , f 4 , f 3 , f 5 } 0.34330.95830.33420.39850.23500.3148
LRFS { f 2 , f 7 , f 8 , f 3 , f 1 , f 4 , f 5 , f 6 } 0.36250.92920.35500.41910.28830.3829
MDMR { f 2 , f 7 , f 3 , f 8 , f 4 , f 5 , f 1 , f 6 } 0.33420.93330.35330.42610.36170.4646
PMU { f 8 , f 7 , f 1 , f 6 , f 2 , f 4 , f 5 , f 3 } 0.37580.98750.27330.33950.19170.2823
SCLS { f 2 , f 7 , f 3 , f 4 , f 1 , f 6 , f 8 , f 5 } 0.34080.95830.35170.43210.33750.4220
Table 4. Description of data sets.
Table 4. Description of data sets.
No.Data Set#Instances#Features#Labels#Training#Test
1medical978144945333645
2scene2407294612111196
3Enron17021001531123579
4Arts50004622620003000
5Business50004383020003000
6Education50005503320003000
7Entertain50006402120003000
8Health50006123220003000
9Recreation50006062220003000
10Reference50007933320003000
11Science50007434020003000
12Social500010473920003000
Table 5. Experimental results of multi-label feature selection methods in terms of Hamming Loss (HL) (mean ± std).
Table 5. Experimental results of multi-label feature selection methods in terms of Hamming Loss (HL) (mean ± std).
Data setMCMFSPPT + MIPPT + CHIMIFSD2FMDMRPMUSCLSLRFS
medical0.015 ± 0.0010.018 ± 0.0010.017 ± 0.0020.017 ± 0.0020.02 ± 0.0010.018 ± 0.0010.02 ± 0.0010.023 ± 00.018 ± 0.001
scene0.135 ± 0.0110.167 ± 0.0060.167 ± 0.0070.17 ± 0.010.149 ± 0.0060.145 ± 0.0070.147 ± 0.0070.173 ± 0.0030.142 ± 0.01
enron0.051 ± 0.0020.053 ± 0.0020.059 ± 0.0010.057 ± 0.0010.052 ± 0.0010.053 ± 0.0030.052 ± 0.0010.053 ± 0.0010.055 ± 0.003
Arts0.060 ± 0.0010.062 ± 0.0010.062 ± 0.0010.061 ± 0.0010.064 ± 0.0010.061 ± 0.0010.064 ± 0.0010.063 ± 0.0010.061 ± 0.001
Business0.029 ± 0.0000.029 ± 0.0010.029 ± 0.0000.028 ± 0.0000.029 ± 0.0010.029 ± 0.0010.029 ± 0.0000.029 ± 00.029 ± 0.001
Education0.042 ± 0.0010.043 ± 0.0010.043 ± 0.0010.044 ± 0.0010.044 ± 0.0010.043 ± 0.0010.045 ± 0.0010.044 ± 0.0010.043 ± 0.001
Entertain0.061 ± 0.0010.064 ± 0.0010.065 ± 0.0010.066 ± 0.0010.066 ± 0.0010.063 ± 0.0020.067 ± 0.0010.066 ± 0.0010.063 ± 0.001
Health0.044 ± 0.0010.046 ± 0.0010.045 ± 0.0020.05 ± 0.0010.048 ± 0.0010.045 ± 0.0010.049 ± 0.0010.049 ± 0.0010.045 ± 0.001
Recreation0.061 ± 0.0010.062 ± 0.0010.062 ± 0.0010.062 ± 0.0010.062 ± 0.0010.062 ± 0.0010.065 ± 0.0010.064 ± 0.0010.061 ± 0.001
Reference0.031 ± 0.0010.032 ± 0.0010.032 ± 0.0010.031 ± 0.0010.032 ± 0.0010.031 ± 0.0010.034 ± 0.0010.033 ± 00.031 ± 0.001
Science0.035 ± 0.0010.036 ± 0.0010.036 ± 0.0000.036 ± 0.0000.036 ± 0.0000.035 ± 0.0000.036 ± 0.0000.036 ± 0.0000.035 ± 0.001
Social0.026 ± 0.0010.028 ± 0.0010.03 ± 0.0010.032 ± 0.0010.03 ± 0.0010.028 ± 0.0010.031 ± 00.029 ± 0.0010.027 ± 0.001
Average0.0490.0530.0540.0550.0530.0510.0530.0550.051
Table 6. Experimental results of multi-label feature selection methods in terms of Zero-One Loss (ZOL) (mean ± std).
Table 6. Experimental results of multi-label feature selection methods in terms of Zero-One Loss (ZOL) (mean ± std).
Data setMCMFSPPT + MIPPT + CHIMIFSD2FMDMRPMUSCLSLRFS
medical0.50 ± 0.050.59 ± 0.050.55 ± 0.060.55 ± 0.080.66 ± 0.040.58 ± 0.040.66 ± 0.040.83 ± 0.010.58 ± 0.04
scene0.57 ± 0.080.78 ± 0.080.8 ± 0.090.83 ± 0.120.61 ± 0.060.61 ± 0.070.61 ± 0.070.74 ± 0.040.6 ± 0.08
enron0.89 ± 0.020.9 ± 0.030.98 ± 00.98 ± 0.010.9 ± 0.020.91 ± 0.030.9 ± 0.030.94 ± 0.030.93 ± 0.04
Arts0.90 ± 0.030.93 ± 0.030.95 ± 0.020.92 ± 0.030.95 ± 0.010.92 ± 0.030.97 ± 0.020.95 ± 0.010.92 ± 0.03
Business0.47 ± 0.010.48 ± 0.010.47 ± 0.010.47 ± 0.010.48 ± 0.010.47 ± 0.010.48 ± 0.010.48 ± 0.010.47 ± 0.01
Education0.88 ± 0.030.91 ± 0.020.94 ± 0.020.95 ± 0.030.95 ± 0.010.9 ± 0.020.95 ± 0.010.93 ± 0.010.9 ± 0.02
Entertain0.82 ± 0.040.87 ± 0.040.9 ± 0.030.93 ± 0.030.91 ± 0.010.85 ± 0.030.94 ± 0.010.9 ± 0.010.86 ± 0.03
Health0.66 ± 0.040.73 ± 0.060.67 ± 0.010.8 ± 0.090.77 ± 0.050.71 ± 0.050.77 ± 0.050.74 ± 0.040.71 ± 0.05
Recreation0.86 ± 0.020.89 ± 0.020.89 ± 0.020.88 ± 0.030.92 ± 0.010.87 ± 0.020.97 ± 0.010.95 ± 0.010.88 ± 0.02
Reference0.74 ± 0.080.74 ± 0.080.71 ± 0.130.78 ± 0.070.8 ± 0.040.76 ± 0.060.81 ± 0.050.83 ± 0.040.76 ± 0.06
Science0.91 ± 0.020.94 ± 0.010.96 ± 0.010.93 ± 0.030.97 ± 0.010.94 ± 0.010.98 ± 0.010.95 ± 0.010.94 ± 0.01
Social0.70 ± 0.050.72 ± 0.050.76 ± 0.130.88 ± 0.090.73 ± 0.090.72 ± 0.050.78 ± 0.070.74 ± 0.040.72 ± 0.05
Average0.740.790.800.820.800.770.820.830.77
Table 7. Experimental results of multi-label feature selection methods in terms of Macro-F1 (mean ± std) using the Support Vector Machine (SVM) classifier.
Table 7. Experimental results of multi-label feature selection methods in terms of Macro-F1 (mean ± std) using the Support Vector Machine (SVM) classifier.
Data setMCMFSPPT + MIPPT + CHIMIFSD2FMDMRPMUSCLSLRFS
medical0.35 ± 0.060.25 ± 0.050.26 ± 0.040.22 ± 0.050.19 ± 0.050.32 ± 0.070.19 ± 0.060.08 ± 0.010.32 ± 0.07
scene0.48 ± 0.090.22 ± 0.090.21 ± 0.10.21 ± 0.150.46 ± 0.080.43 ± 0.070.47 ± 0.090.26 ± 0.050.44 ± 0.08
enron0.12 ± 0.030.1 ± 0.030.07 ± 0.020.07 ± 0.020.13 ± 0.040.11 ± 0.030.13 ± 0.050.12 ± 0.030.1 ± 0.03
Arts0.09 ± 0.030.06 ± 0.020.07 ± 0.020.07 ± 0.020.03 ± 0.000.08 ± 0.030.01 ± 0.010.03 ± 0.000.07 ± 0.02
Business0.05 ± 0.000.05 ± 0.000.05 ± 0.000.04 ± 0.000.05 ± 0.000.05 ± 0.000.03 ± 0.000.04 ± 0.000.05 ± 0.00
Education0.07 ± 0.010.06 ± 0.010.05 ± 0.010.03 ± 0.020.05 ± 0.010.06 ± 0.010.03 ± 0.010.04 ± 0.010.06 ± 0.01
Entertain0.13 ± 0.030.11 ± 0.030.09 ± 0.020.06 ± 0.020.08 ± 0.010.12 ± 0.020.05 ± 0.000.07 ± 0.010.12 ± 0.02
Health0.15 ± 0.030.13 ± 0.030.14 ± 0.030.06 ± 0.030.09 ± 0.010.14 ± 0.030.08 ± 0.010.09 ± 0.010.14 ± 0.03
Recreation0.11 ± 0.020.1 ± 0.020.1 ± 0.020.09 ± 0.030.08 ± 0.010.11 ± 0.020.03 ± 0.000.04 ± 0.000.11 ± 0.02
Reference0.07 ± 0.010.07 ± 0.010.07 ± 0.020.06 ± 0.020.04 ± 0.000.07 ± 0.010.03 ± 0.010.02 ± 0.000.07 ± 0.01
Science0.07 ± 0.020.05 ± 0.020.05 ± 0.010.04 ± 0.020.02 ± 0.000.05 ± 0.020.01 ± 0.010.02 ± 0.000.05 ± 0.02
Social0.11 ± 0.030.09 ± 0.020.09 ± 0.020.05 ± 0.030.07 ± 0.010.1 ± 0.030.05 ± 0.010.05 ± 0.010.1 ± 0.03
Average0.150.110.100.090.110.140.090.070.14
Table 8. Experimental results of multi-label feature selection methods in terms of Macro-F1 (mean ± std) using the 3-Nearest Neighbors (3NN) classifier.
Table 8. Experimental results of multi-label feature selection methods in terms of Macro-F1 (mean ± std) using the 3-Nearest Neighbors (3NN) classifier.
Data setMCMFSPPT + MIPPT + CHIMIFSD2FMDMRPMUSCLSLRFS
medical0.25 ± 0.040.16 ± 0.030.19 ± 0.020.16 ± 0.020.12 ± 0.020.19 ± 0.030.11 ± 0.020.06 ± 0.010.19 ± 0.03
scene0.54 ± 0.070.37 ± 0.080.36 ± 0.080.29 ± 0.140.49 ± 0.050.51 ± 0.060.49 ± 0.070.37 ± 0.030.53 ± 0.07
enron0.13 ± 0.020.12 ± 0.020.07 ± 0.010.09 ± 0.010.12 ± 0.010.12 ± 0.020.12 ± 0.020.11 ± 0.010.11 ± 0.02
Arts0.11 ± 0.030.08 ± 0.020.1 ± 0.030.08 ± 0.030.06 ± 0.010.1 ± 0.020.06 ± 0.010.07 ± 0.020.1 ± 0.02
Business0.1 ± 0.010.08 ± 0.010.09 ± 0.010.09 ± 0.020.07 ± 0.010.09 ± 0.010.05 ± 0.010.07 ± 0.010.08 ± 0.01
Education0.09 ± 0.020.08 ± 0.020.08 ± 0.020.04 ± 0.020.06 ± 0.010.07 ± 0.010.06 ± 0.010.06 ± 0.010.07 ± 0.01
Entertain0.14 ± 0.030.13 ± 0.020.11 ± 0.020.08 ± 0.020.11 ± 0.010.13 ± 0.020.08 ± 0.010.09 ± 0.010.14 ± 0.02
Health0.14 ± 0.030.11 ± 0.020.12 ± 0.030.05 ± 0.030.09 ± 0.010.12 ± 0.020.09 ± 0.010.09 ± 0.010.12 ± 0.02
Recreation0.13 ± 0.020.1 ± 0.010.11 ± 0.020.12 ± 0.030.08 ± 0.010.12 ± 0.020.05 ± 0.010.07 ± 0.010.12 ± 0.02
Reference0.08 ± 0.010.07 ± 0.010.08 ± 0.020.07 ± 0.010.04 ± 00.07 ± 0.010.03 ± 0.010.04 ± 0.010.07 ± 0.01
Science0.08 ± 0.020.05 ± 0.010.07 ± 0.010.06 ± 0.020.04 ± 0.010.07 ± 0.020.03 ± 0.010.03 ± 00.06 ± 0.01
Social0.12 ± 0.020.08 ± 0.010.1 ± 0.010.07 ± 0.030.06 ± 0.010.09 ± 0.010.05 ± 0.010.05 ± 00.09 ± 0.01
Average0.160.120.120.100.110.140.100.090.14
Table 9. Experimental results of multi-label feature selection methods in terms of Micro-F1 (mean ± std) using the SVM classifier.
Table 9. Experimental results of multi-label feature selection methods in terms of Micro-F1 (mean ± std) using the SVM classifier.
Data setMCMFSPPT + MIPPT + CHIMIFSD2FMDMRPMUSCLSLRFS
medical0.79 ± 0.050.73 ± 0.050.74 ± 0.070.71 ± 0.110.63 ± 0.070.76 ± 0.050.63 ± 0.080.37 ± 0.010.76 ± 0.05
scene0.50 ± 0.080.25 ± 0.10.24 ± 0.110.24 ± 0.160.48 ± 0.070.46 ± 0.070.49 ± 0.080.3 ± 0.050.47 ± 0.07
enron0.51 ± 0.030.47 ± 0.040.35 ± 0.020.37 ± 0.030.51 ± 0.030.47 ± 0.050.5 ± 0.040.49 ± 0.030.45 ± 0.06
Arts0.18 ± 0.050.14 ± 0.050.12 ± 0.040.17 ± 0.050.08 ± 0.010.17 ± 0.050.03 ± 0.020.07 ± 0.020.16 ± 0.05
Business0.68 ± 0.000.68 ± 0.000.68 ± 0.000.67 ± 0.000.67 ± 0.000.68 ± 0.000.67 ± 0.000.67 ± 00.68 ± 0.00
Education0.23 ± 0.050.2 ± 0.040.13 ± 0.040.12 ± 0.060.12 ± 0.020.21 ± 0.050.08 ± 0.010.14 ± 0.020.21 ± 0.04
Entertain0.27 ± 0.060.23 ± 0.060.17 ± 0.050.11 ± 0.050.16 ± 0.010.26 ± 0.060.1 ± 0.010.15 ± 0.020.25 ± 0.06
Health0.50 ± 0.020.45 ± 0.070.47 ± 0.030.39 ± 0.050.42 ± 0.010.47 ± 0.040.39 ± 0.030.41 ± 00.48 ± 0.03
Recreation0.20 ± 0.040.19 ± 0.030.17 ± 0.040.18 ± 0.050.14 ± 0.020.2 ± 0.040.04 ± 00.07 ± 0.010.2 ± 0.04
Reference0.32 ± 0.040.35 ± 0.070.35 ± 0.140.33 ± 0.10.31 ± 0.040.34 ± 0.060.27 ± 0.050.26 ± 0.040.34 ± 0.06
Science0.15 ± 0.040.12 ± 0.030.09 ± 0.030.11 ± 0.050.05 ± 0.010.13 ± 0.030.02 ± 0.020.06 ± 0.010.13 ± 0.03
Social0.45 ± 0.080.42 ± 0.070.38 ± 0.140.2 ± 0.120.4 ± 0.070.43 ± 0.070.31 ± 0.070.38 ± 0.050.43 ± 0.07
Average0.400.350.330.300.330.380.290.280.38
Table 10. Experimental results of multi-label feature selection methods in terms of Micro-F1 (mean ± std) using the 3NN classifier.
Table 10. Experimental results of multi-label feature selection methods in terms of Micro-F1 (mean ± std) using the 3NN classifier.
Data setMCMFSPPT + MIPPT + CHIMIFSD2FMDMRPMUSCLSLRFS
medical0.69 ± 0.040.62 ± 0.040.64 ± 0.060.61 ± 0.10.53 ± 0.040.64 ± 0.030.52 ± 0.040.35 ± 0.010.64 ± 0.03
scene0.55 ± 0.060.39 ± 0.060.38 ± 0.060.34 ± 0.110.49 ± 0.040.52 ± 0.050.5 ± 0.050.38 ± 0.020.54 ± 0.05
enron0.49 ± 0.020.45 ± 0.010.34 ± 0.030.41 ± 0.020.47 ± 0.030.44 ± 0.040.47 ± 0.020.44 ± 0.030.42 ± 0.05
Arts0.26 ± 0.050.17 ± 0.050.24 ± 0.040.18 ± 0.050.15 ± 0.030.25 ± 0.040.14 ± 0.030.17 ± 0.030.25 ± 0.04
Business0.67 ± 0.010.67 ± 0.000.66 ± 0.010.65 ± 0.080.66 ± 0.000.67 ± 0.010.65 ± 0.040.60 ± 0.120.67 ± 0.01
Education0.26 ± 0.030.24 ± 0.040.28 ± 0.040.16 ± 0.060.19 ± 0.030.23 ± 0.030.18 ± 0.040.19 ± 0.030.23 ± 0.03
Entertain0.27 ± 0.050.28 ± 0.050.21 ± 0.050.22 ± 0.080.24 ± 0.030.26 ± 0.040.22 ± 0.050.22 ± 0.030.27 ± 0.03
Health0.37 ± 0.090.38 ± 0.070.37 ± 0.140.2 ± 0.070.37 ± 0.050.38 ± 0.060.36 ± 0.040.37 ± 0.060.38 ± 0.05
Recreation0.25 ± 0.030.19 ± 0.020.21 ± 0.040.23 ± 0.050.16 ± 0.020.23 ± 0.040.09 ± 0.020.12 ± 0.020.23 ± 0.03
Reference0.46 ± 0.040.41 ± 0.050.39 ± 0.130.35 ± 0.090.36 ± 0.050.43 ± 0.050.35 ± 0.040.29 ± 0.050.43 ± 0.04
Science0.18 ± 0.040.17 ± 0.030.12 ± 0.030.17 ± 0.020.12 ± 0.020.16 ± 0.030.1 ± 0.020.15 ± 0.030.16 ± 0.03
Social0.46 ± 0.050.4 ± 0.060.44 ± 0.10.39 ± 0.050.39 ± 0.050.42 ± 0.060.36 ± 0.050.37 ± 0.040.41 ± 0.05
Average0.410.360.360.330.340.380.330.300.38
Table 11. Running time (seconds).
Table 11. Running time (seconds).
Data SetMCMFSD2FMDMRPMUSCLSLRFSPPT + MIPPT + CHIMIFS
medical142.210,698.59910.011521.638.26961.91.28.630.3
scene23.5244.1246.1257.05.5147.00.41.847.4
enron267.322,164.020,583.025749.693.616,631.90.86.441.2
Arts111.94355.84185.45002.239.63421.21.04.135.5
Business107.44536.14361.85354.041.54079.81.03.141.9
Education161.28397.37700.09037.259.16323.11.24.399.3
Entertain194.46809.96539.27386.654.34519.11.45.098.7
Health195.19376.09086.810,797.168.47523.41.34.817.2
Recreation183.46394.26127.67016.851.74369.01.34.639.5
Reference313.516,269.815,637.318,483.696.211,685.11.85.757.5
Science291.417,419.916,665.220,137.1102.714,344.91.66.660.0
Social553.834,108.833,195.637,246.0156.123,754.52.89.043.4

Share and Cite

MDPI and ACS Style

Zhang, P.; Gao, W.; Hu, J.; Li, Y. Multi-Label Feature Selection Based on High-Order Label Correlation Assumption. Entropy 2020, 22, 797. https://0-doi-org.brum.beds.ac.uk/10.3390/e22070797

AMA Style

Zhang P, Gao W, Hu J, Li Y. Multi-Label Feature Selection Based on High-Order Label Correlation Assumption. Entropy. 2020; 22(7):797. https://0-doi-org.brum.beds.ac.uk/10.3390/e22070797

Chicago/Turabian Style

Zhang, Ping, Wanfu Gao, Juncheng Hu, and Yonghao Li. 2020. "Multi-Label Feature Selection Based on High-Order Label Correlation Assumption" Entropy 22, no. 7: 797. https://0-doi-org.brum.beds.ac.uk/10.3390/e22070797

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop