Next Article in Journal
Electroosmosis-Optimized Thermal Model for Peristaltic Transportation of Thermally Radiative Magnetized Liquid with Nonlinear Convection
Next Article in Special Issue
A Novel Method for Fault Diagnosis of Rotating Machinery
Previous Article in Journal
An Effective Synchronization Approach to Stability Analysis for Chaotic Generalized Lotka–Volterra Biological Models Using Active and Parameter Identification Methods
Previous Article in Special Issue
Stock Index Prediction Based on Time Series Decomposition and Hybrid Model
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Preference-Driven Classification Measure

1
Department of Machine Learning, University of Economics in Katowice, 1 Maja 50, 40-287 Katowice, Poland
2
Department of Knowledge Engineering, University of Economics in Katowice, 1 Maja 50, 40-287 Katowice, Poland
*
Author to whom correspondence should be addressed.
Submission received: 7 February 2022 / Revised: 4 April 2022 / Accepted: 8 April 2022 / Published: 10 April 2022
(This article belongs to the Special Issue Entropy in Real-World Datasets and Its Impact on Machine Learning)

Abstract

:
Classification is one of the main problems of machine learning, and assessing the quality of classification is one of the most topical tasks, all the more difficult as it depends on many factors. Many different measures have been proposed to assess the quality of the classification, often depending on the application of a specific classifier. However, in most cases, these measures are focused on binary classification, and for the problem of many decision classes, they are significantly simplified. Due to the increasing scope of classification applications, there is a growing need to select a classifier appropriate to the situation, including more complex data sets with multiple decision classes. This paper aims to propose a new measure of classifier quality assessment (called the preference-driven measure, abbreviated p-d), regardless of the number of classes, with the possibility of establishing the relative importance of each class. Furthermore, we propose a solution in which the classifier’s assessment can be adapted to the analyzed problem using a vector of preferences. To visualize the operation of the proposed measure, we present it first on an example involving two decision classes and then test its operation on real, multi-class data sets. Additionally, in this case, we demonstrate how to adjust the assessment to the user’s preferences. The results obtained allow us to confirm that the use of a preference-driven measure indicates that other classifiers are better to use according to preferences, particularly as opposed to the classical measures of classification quality assessment.

1. Introduction

Classification continues to be one of the most important subjects in machine learning. Despite this, we still lack a general measure of quality independent of the specific characteristics of the data set. Moreover, in the situations where there is a necessity to involve the human decision maker in the classification process, we are forced to switch between different measures. Among the general ones, there are accuracy, precision or recall, and others that are data-dependent. Choosing the right (optimal) one is especially important because choosing a particular classification method depends heavily on the calculated quality measures.
Moreover, there is no single best classification measure that effectively identifies the method suitable for every task. Classification algorithms/methods have many characteristics. Consequently, there are many measures of classification because there is no single measure covering all the characteristics simultaneously [1]. Thus, finding an appropriate classification measure for a specific task is difficult and requires answering the question of what conditions, in specific circumstances, must be met by the measure.
One of the ways to bypass the problem of unambiguous assessment of the classifier’s quality and selecting one best-suited classifier is ensemble and hybrid methods, which simultaneously use many different algorithms to perform a specific task. Within this approach, we can point out the homogeneous and heterogeneous solutions [2]. In the first group, we can find methods allowing us to create large groups of classifiers belonging to the same category (or even classifiers generated with the same method, but with different starting parameters), which allow the classification process simultaneously. The heterogeneous approach involves using a large variety of classifiers, for which the main advantage is the diversity of obtained results. Thus, the basic idea of these methods is the idea of collective decision making [3].
One should note that the above approach based on the ensemble methods allows for more robust classifier selection. However, it still leaves the decision maker with the problem of estimating the classification quality. In medicine, military or finance, the well-known accuracy measure seems to derive unsatisfactory results and present limited usability [4,5]. On the other hand, measures such as recall or precision are directed towards the binary classification problem. Most of the proposed measures are directly connected to the confusion matrix and related absolutely to the numerical outcome of the classification [6]. In most cases, it is strongly needed, but it makes the whole process independent of the user’s preferences. In the decision-making context, taking them into account may be vital to make the process effective and at the same time maintain the sovereignty of the decision maker.
Incorporating users in the process of preparing a machine learning solution is an essential element of the entire procedure, the subject of many studies, and can take various forms [7,8]. One of the goals of the actions taken is to help the user to choose a suitable classifier. Most often, this task comes down to comparing simple measures of classification quality, which is usually carried out by trial and error, and yet can be unreliable [9]. This task becomes even more complicated if individual users’ preferences are to be taken into account. This applies especially to issues of a managerial nature, but more generally wherever a human being to some extent participates in the decision-making process. In practice, this applies to all issues except physical or technical ones, where only objective laws are in force [10]. The main intention of introducing the new measure is to make this stage of a research procedure more methodical. To the best of our knowledge, there are no clear guidelines for taking into account the parameters in the learning process and the selection of a classifier in conjunction with individual preferences. The next issue is the systematic classification into the following application areas, thus expanding the group using machine learning methods. Finally, some users need a tool to control the process of selecting a classifier for their own needs, which are more complex and related to many classes. For such users, a measure that allows them to simply, directly, and methodically include their preferences in selecting and training the classifier would be very beneficial.
To cope with the above drawbacks, and at the same time maintain the role of the decision maker in the process, we propose the idea of a new measure in which their preferences are vital to the importance of individual classes. We aim to propose a measure that balances a thorough analysis of the classifier’s performance and the selection of the classifier that performs best under the conditions (preferences) specified by the user.
The main intention of introducing the new measure is to make this stage of the research procedure more methodical. Users need a tool with which they will be able to select a classifier for their own needs, which are more complex and related to many classes. A measure that allows them to directly and methodically include their preferences in selecting and training the classifier would be very beneficial for such users.
We undertook to propose such a measure because, to the best of our knowledge, there is no reasonable alternative where, for any number of decision classes, it is possible to aggregate the quality of classification depending on the weight for a particular decision class. First, our solution was discussed in detail on prepared examples for two decision classes. It was then tested on real-world data sets and re-examined for a more significant number of decision classes that occurred in these data sets.
This article is organized as follows. Section 1 introduces the subject of this article. Section 2 provides an overview of the work related to the classification, particularly the measures for assessing the quality of the classification. In Section 3, we describe the classification problem and the classification quality assessment measures based on the error matrix for binary and multi-class classification. In Section 4, we present a new measure for classification, in which it will be possible to control preferences. In Section 5, we present the analysis of our research on real data sets. Finally, in Section 6 and Section 7, we discuss the results of the experiments and end with general remarks on this work and available directions for future research.

2. Related Works

Evaluating the classification performance is a difficult task, and the discussion on this topic arose from the beginning of work on automatic classification. The initial set of five measures (sensitivity, specificity, efficiency, positive and negative predictive value) was rapidly expanded [11]. The most often used measure of classification performance is accuracy. However, it is not the only measure of the quality of predictive models. Despite optimizing the classification error rate, high-accuracy models may fail to capture crucial information transfer in the classification task [12,13]. Despite the simplicity and intuitive interpretation, there are many reasons and situations in which accuracy should not be used [14]. Instead, the authors advocate for using Cohen’s kappa as a better meter for measuring classifiers’ own merits than accuracy. Moreover, [15] indicates that the most frequently used measures, which focus on correctly classified cases (precision, recall, or F-score), do not meet the needs of various decision-making situations, especially when more than one class is essential. The authors advocate for using three other measures—Youden’s index, likelihood, and discriminant power—because they combine sensitivity and specificity and their complements.
While most studies concern binary classification, in [16], the authors focused on multi-class classification problems. The authors showed that the extension of measures to the classification of many classes is associated with averaging the results achieved for individual classes in most cases. Finally, in [17], the authors point out some shortcomings of the accuracy measure and list five conditions that the newly constructed discriminator metric should meet.
To solve the dilemma related to the choice of a measure for a given problem, a list of desired features of an ideal measure and analysis of the most known measures was proposed in [1]. More importantly, they proved that it is impossible to satisfy them simultaneously. They also proposed a new family of measures (Generalized Means) that meet all desirable properties except one, and a new measure called Symmetric Balanced Accuracy.
A comparative analysis and taxonomy of the quality of classification measures have been the subject of many studies. For instance [18], in their experimental comparison conducted on 30 data sets, proposed dividing the performance measures into three categories:
  • Measures based on a threshold, such as accuracy, modified accuracy measure, F-measure, or Kappa statistic, which are used to minimize the number of errors in the model. They are based directly on a confusion matrix, and they are widely used in many classification tasks. One should note that the overall efficiency of these measures is strictly related to the quality of the data. However, some measures, such as accuracy, can be less effective in the case of unbalanced data sets.
  • Measures based on a probabilistic approach to understanding error, i.e., measuring the deviation or entropy information between the actual and predicted probability, such as mean absolute error, mean square error (Brier score), LogLoss (cross-entropy). These measures are useful in measuring the probability of selecting the wrong class, which is essential in ensemble methods or for a committee of classifiers.
  • Measures based on the model’s ability to correctly rank cases include ROC, AUC, the Youden index, precision–recall curve, Kolmogorov–Smirnov, or lift chart. They are helpful when indicating the best n occurrences in a data set or when good class separation is needed. They are widely used in recommendation systems, design marketing campaigns, fraud detection, spam filtering, and more.
In a survey [19], the authors grouped the measures depending on the type of outcome on which a given measure is focused (correct or incorrect outcome). In turn, [20] presents an in-depth analysis of over twenty performance measures used in different classification tasks in the context of changes made to a confusion matrix and their relations with particular measures (measure invariance). In contrast, a comparative study of two or more classifiers based on statistical tests was presented in [21]. A comprehensive analysis of the methods and measures of classification assessment is also included in [22], where the relationships between all measures calculated based on the confusion matrix are shown. Finally, a different approach to the analysis of performance measures was presented in [23]. First, the authors grouped classification measures according to classification difficulty, which they defined in relation to a distance between the boundary lines and each correctly classified case. The authors later developed their idea and proposed an instance-based measure for calculating the performance of classification from the perspective of instances, called degree of credibility [24].
The set of measures is constantly growing. For instance, in [25] was proposed a measure that compares classifiers, which combines three measures from different groups: Matthews correlation coefficient as a measure, which is calculated from both true and false positives and negatives, and AUC (area under the curve), derived from ROC and accuracy. To overcome the shortcomings of the accuracy measure in evaluating multi-class classifiers and to improve the quality of classifiers, in [26], the authors proposed a metric based on the combined accuracy and dispersion values. They also showed experimentally that this two-dimensional metric is particularly suitable in complex, unbalanced data sets and with many classes.
The new measures are also proposed to supplement the already used measures that work better for specific tasks. For example, [27], as an alternative to measures used in medical diagnostics, which use only part of the values from the error matrix, define the measure A Q M , which takes into account all values from the confusion matrix. Similarly, in [28], for image analysis, a new measure of classification performance called robust-and-balanced accuracy was introduced. It aims to connect balanced accuracy with measures of variations. In another proposition, to improve face recognition processes, a new classification measure, called the volume measure, based on the volume of the matrix, was proposed [29]. In turn, a measure dedicated to the analysis of imbalanced data sets based on the harmonic mean of recall and selectivity was proposed in [30].
Existing measures are also modified. For example, in [31], the F * measure was proposed as a modification of the F-score, towards the more straightforward interpretation of this measure. The above short review shows that the issue of classification measures is constantly under the attention of researchers. Moreover, for specific applications, new measures are created that are better suited to the requirements of the domain or user preferences.

3. Classification and Quality Evaluation

Formally, the classification problem Q can be solved using empirical experience W, while the quality of the solution is estimated by the quality measure Y. The value of Y should be increasing, while the experience W rises as well [32].
In machine learning, classification refers to the prediction problem of determining the class to which samples from a data set will be assigned. A classifier algorithm (often shortened to “a classifier”) must be provided with training data with labeled classes. Then, the classifier can predict classes for new test data based on the training data. This approach is called supervised machine learning, and classification is one example of such a method. The training set is selected as a subset of the whole data set. A typical approach is to divide the known examples into a train and test set, following some general principles about the ratio of the two. Eventually, the test set includes a far smaller number of samples than the training set (preferably, the test set and training set should be disjoint). Test set is used to evaluate classification quality. Every object (also called a sample or observation) from the training data is assigned some predefined label (decision class). The idea is to build such a classifier, which assigns the proper labels for the objects. In contrast, the evaluation is performed on the training set, where the difference between the assigned and the actual label is estimated.
Classification algorithms for a prediction problem are evaluated based on performance. Multiple measures for assessing classification quality can be used depending on the situation. One of the most commonly used measures is accuracy, which determines how many samples from the entire data set were correctly classified into the appropriate classes. Often, other measures are used in addition to accuracy that more accurately assess the classifier’s performance. Such measures primarily include precision and recall. Unfortunately, there are many problems in which classical measures are insufficient, and it is necessary to look for new solutions.
Let us consider a set of all available samples (called the universe of objects) X, which will include n number of objects:
X = { x 1 , x 2 , , x i , , x n } ,
A single observation x i will be described by a finite set of attributes and the decision attribute:
a 1 , a 2 , , a m ,
where a j A j , j = 1 , , m . In this context, A j denotes the domain of the j-th attribute, while features a 1 , a 2 , , a m create a feature space A 1 × A 2 × × A m .
There are no general restrictions related to the values of attributes, which can be quantitative or categorical. However, one should note that preprocessing allows for discretizing selected quantitative attributes. The same procedure can be applied to the decision classes, where many labels can be merged into fewer during the discretization process (trade-off between the quality of classification and classification speed). Thus, the single sample can be described as follows:
x i = ( V i , c k ) , v i j A j , c k { 1 , , C } ,
where V i = [ v i 1 , , v i m ] is a vector in an m-dimensional feature space, v i j is the value of attribute a j for observation (sample) x i , and c i is the class label (also called the decision class) of this object. Thus, the universe X can be formally described as:
X : { ( V i , c i ) } i = 1 n .
Hence, the classification problem can generally be understood as the assignment problem, where every element x i from the universe X should have the decision class c i assigned. Eventually, we end with the classifier capable of assigning the newly arrived objects from the test set into proper decision classes. However, even in the case of the binary classification problem, the above is not trivial. It can be challenging to achieve when, for example, we observe unbalanced data (i.e., the situation in which most of the objects from universe X are assigned to a single class). Therefore, many different estimation methods were proposed (both to binary and multiple-class classification problems) to cope with this. Next, this section discusses various classification measures capable of dealing with both mentioned classification problems.

3.1. Quality Assessment and Binary Confusion Matrix

A confusion matrix is among the most popular tools used in the process of the validation of the quality of performed classification. The confusion matrix for a binary decision class is defined as a table with different combinations of predicted and actual values related to the decision classes [33,34]. The rows contain existing classes, while the columns contain predicted classes. The diagonal of the confusion matrix includes the correctly classified samples for both classes, while the off-diagonal cells represent the errors (misclassified samples for both decision classes). The confusion matrix represents the errors that the classifier makes and shows the type of these errors. It represents a detailed breakdown of the answers considering the number of correct and incorrect classifications for both classes. It is imperative when the cost of misclassification is different for these classes and when the size of the classes is different [35].
For the binary classification (e.g., classification of emails and spam or sick and healthy patients), see Table 1, where the prediction of a positive class (labeled 1) and a negative class (labeled 2) can be uniquely determined. Such a situation indicates whether the classifier is more likely to incur an error by assigning a positive class as a negative class or vice versa. In Table 1, the confusion matrix for the binary classifier is shown, where:
  • T P (true positive)–samples classified as predicted class 1, which are samples of actual class 1;
  • F N (false negative)–samples classified as predicted class 2, which are samples of actual class 1;
  • F P (false positive)–samples classified as predicted class 1, which are samples of actual class 2;
  • T N (true negative)–samples classified as predicted class 2, which are samples of actual class 2.
The quality of classifiers is measured based on a confusion matrix (as shown in Table 1). Measures based on the confusion matrix include, but are not limited to, accuracy, precision, recall, and F-score (also called F1, which is the measure F β for which β = 1 ). In addition, the Matthews correlation coefficient (MCC) measure and BalancedAccuracy are also often used for binary classification.
The accuracy measure indicates how often a classifier makes a correct prediction. It is the ratio of the number of accurate predictions to the total number of predictions and is calculated based on Formula (5).
a c c u r a c y _ b i n a r y = T P + T N T P + F P + F N + T N
Precision determines how many samples, out of all those classified as positive, are samples of the positive class. Precision is expressed by Formula (6).
p r e c i s i o n _ b i n a r y = T P T P + F P
Recall is used to determine how many samples belonging to a positive class were classified as positive by the classifier. Recall is determined by Formula (7).
r e c a l l _ b i n a r y = T P T P + F N
For binary classification, the measure F β is defined as the harmonic mean of precision and recall, where additional precision or recall weights are used to obtain more accurate results. By setting the value of β , it is possible to control the effect of recall weight with respect to precision. F β is specified by Formula (8), where β is the number of times the recall is as important as the precision [36]. The F β value ranges from 0 to 1 (with 0 being the worst value and 1 being the optimal value). The most common value for β is 1, which simply means measure F1 (Equation (9)). Other frequently used values are 2 and 0.5. In the case of 2, recall weight is greater than precision; however, in the case of 0.5, recall weight is less than precision. The F β measure is based on the Van Rijsbergen measure of effectiveness [37].
F β = ( 1 + β 2 ) × p r e c i s i o n _ b i n a r y × r e c a l l _ b i n a r y β 2 × p r e c i s i o n _ b i n a r y + r e c a l l _ b i n a r y
The F 1 measure is the instance of the measure F β (Equation (8)), for which β = 1 . The F 1 measure is defined as the harmonic mean of precision and recall [38]. Therefore, F 1 will only take on a high value if both of its components reach a high value. The F 1 measure often replaces precision when class counts are unbalanced [39]. For example, if 97% of the data belong to class 1 and only 3% belong to class 2, then classifying all observations as class 1 would yield a misleading accuracy of 97%. The F 1 measure is based on precision and recall, and is thus robust to such distortions. The measure is calculated from Formula (9).
F 1 _ b i n a r y = 2 × p r e c i s i o n _ b i n a r y × r e c a l l _ b i n a r y p r e c i s i o n _ b i n a r y + r e c a l l _ b i n a r y = T P T P + 1 2 ( F N + F P )
The Matthews correlation coefficient ( M C C ) measure is often used for unbalanced data [40]. For the precision, recall, and F 1 measures, the T N value is not used, which is very important if we are interested in both classes. Therefore, the Matthews correlation coefficient, calculated based on all terms from the confusion matrix, can be used. The M C C measure is calculated from Formula (10).
M C C _ b i n a r y = T P × T N F P × F N ( T P + F P ) ( T P + F N ) ( T N + F P ) ( T N + F N )
The M C C value ranges within [ 1 1 ] (with −1 equal to the misclassification of all samples, while a value of 1 means that all samples are correctly classified). Therefore, the higher the correlation between actual and predicted values, the better the prediction. Furthermore, for the M C C measure, the two classes have the same importance weight, so when positive classes are swapped with negative classes, the M C C value will be the same, which means that the M C C measure is symmetric [40].
Another measure in binary classification is the BalancedAccuracy measure, which calculates balanced accuracy for anomaly or disease detection, where significant differences in the class size are observed [41]. Overestimated accuracy results can be avoided using the BalancedAccuracy measure for unbalanced classes. For binary classification, the balanced accuracy equals the arithmetic mean of the recall and specificity. The BalancedAccuracy measure is calculated by Formulas (11) or (12).
B a l a n c e d A c c u r a c y _ b i n a r y = r e c a l l _ b i n a r y + s p e c i f i c i t y _ b i n a r y 2
B a l a n c e d A c c u r a c y _ b i n a r y = 1 2 T P T P + F N + T N F P + T N
Specificity is used to determine the number of samples belonging to a negative class that have been classified as negative by the classifier [42]. Specificity is measured by Formula (13).
s p e c i f i c i t y _ b i n a r y = T N F P + T N

3.2. Quality Assessment and Confusion Matrix for Multi-Class Classification

In addition to binary classification, it is common to classify samples into more than two classes. We are dealing with multi-class classification in such a situation—not to be confused with multi-label classification. The difference between multi-class and multi-label classification is that a sample can only be assigned to one class selected from multiple classes for multi-class classification. In contrast, a sample can be assigned to multiple classes for multi-label classification [43].
There is a need to construct an extended confusion matrix for a number of decision classes greater than 2. Thus, the number of rows and the number of columns in such an approach will equal the number of classes. For this purpose, we present the idea of a confusion matrix for C classes, which is shown in Table 2. Similarly, for the binary classification problem (see Table 1), we have listed four terms:
  • T P (true positive) means that samples from the actual class have been classified into the same predicted class—denoted as T P 1 ,   T P 2 , ,   T P i , ,   T P C 1 ,   T P C ;
  • F N (false negative) indicates that samples in the actual class have been classified into other predicted classes. This is the sum of the values of the corresponding row of the real class except for the T P values for that real class—denoted as F N 1 ,   F N 2 , ,   F N i , ,   F N C 1 ,   F N C ;
  • F P (false positive) indicates that samples from other real classes have been classified into the selected predicted class. This is the sum of the values of the corresponding column of the predicted class except for the T P values for the actual class—denoted as F P 1 ,   F P 2 , ,   F P i , ,   F P C 1 ,   F P C ;
  • T N (true negative) indicates that, for the selected real class, samples from other actual classes were classified into predicted classes other than the predicted class corresponding to the chosen real class. For a given real class, it is the sum of the values of all columns and rows except for the row and column values of that real class for which we compute the values—denoted as T N \ { i } or T N \ { i , C } .
To evaluate the quality of multi-class classification, as in binary classification, measures based on the confusion matrix shown in Table 2 were used. Therefore, referring to the measures described in Section 3.1, we wish to present them for the multi-class classification problem.
Accuracy is one of the most commonly used measures in a multi-class classification problem [16] and is calculated according to Formula (14) or (15); when we define the sum of all samples as s, this equation boils down to the form (16). To calculate accuracy, sum all correctly classified samples and then divide by the number of all classified samples. Correctly classified samples are shown in the confusion matrix (see Table 2) on the main diagonal (from upper left corner to lower right corner).
a c c u r a c y = i = 1 c T P i i = 1 c T P i + i = 1 c F P i
a c c u r a c y = i = 1 c T P i i = 1 c T P i + i = 1 c F N i
a c c u r a c y = i = 1 c T P i s
Precision and recall are used for the multi-class approach as well. Two modifications can be distinguished in this case: m a c r o and m i c r o precision or recall [16]. For the m a c r o modification, to calculate the value of the measures for multiple classes, one must count precision and recall for each class separately and then calculate the arithmetic mean of these values. In this way, all classes during multi-class classification have the same validity, regardless of the class count. In multi-class classification, m a c r o _ p r e c i s i o n is calculated by Formula (17) and m a c r o _ r e c a l l is calculated by Formula (18).
m a c r o _ p r e c i s i o n = 1 c i = 1 c T P i T P i + F P i
m a c r o _ r e c a l l = 1 c i = 1 c T P i T P i + F N i
In contrast, for m i c r o _ modification, to count the precision and recall values, one must look at all classes together. In this way, each correctly classified sample into a class is a component of all correctly classified samples. In other words, we calculate T P as the sum of all T P values for individual classes (the sum of the values from the main diagonal). The F P value will be the sum of all values off the main diagonal, equal to the F N value. Therefore, m i c r o _ p r e c i s i o n and m i c r o _ r e c a l l are the same because they are the sum of T P values to all values in the confusion matrix [16]. In multi-class classification, m i c r o _ p r e c i s i o n is calculated by Formula (19) and m i c r o _ r e c a l l is calculated by Formula (20).
m i c r o _ p r e c i s i o n = i = 1 c T P i i = 1 c T P i + i = 1 c F P i
m i c r o _ r e c a l l = i = 1 c T P i i = 1 c T P i + i = 1 c F N i
The F β measure is also used for multi-class classification problems; however, the results of this measure are obtained by macro-averaging or micro-averaging [20]. When we assume that all classes are of the same weight, we use macro-averaging, where two additional methods can be adopted. The first is the calculation of the F β measure from the m a c r o _ p r e c i s i o n (see Equation (17)) and m a c r o _ r e c a l l (see Equation (18)) measures, and the second is the arithmetic mean of the F-beta scores for each class separately, based on the F β measure for binary classification (see Equation (8)). The m a c r o _ F β measure is specified by Formula (21).
m a c r o _ F β = ( 1 + β 2 ) × m a c r o _ p r e c i s i o n × m a c r o _ r e c a l l β 2 × m a c r o _ p r e c i s i o n + m a c r o _ r e c a l l
In contrast, in the classification depending on the frequency of classes, the F β results are calculated by micro-averaging based on m i c r o _ p r e c i s i o n (see Equation (19)) and m i c r o _ r e c a l l (see Equation (20)). The m i c r o _ F β measure is specified by Formula (22). In all of the discussed variants of the F β measure, the beta parameter determines the weight of the reference in relation to the precision. When β < 1 , precision has more weight, and when β > 1 , recall has more weight [20].
m i c r o _ F β = ( 1 + β 2 ) × m i c r o _ p r e c i s i o n × m i c r o _ r e c a l l β 2 × m i c r o _ p r e c i s i o n + m i c r o _ r e c a l l
The F1 measure (the most common instance of F β , where β = 1 ) is calculated based on recall and precision (present in the multi-class classification) [44], and thus can be used in the multi-class classification as well. For a problem where class imbalance does not matter and all classes are equally valid, we use the m a c r o _ modification and apply m a c r o _ p r e c i s i o n (see Equation (17)) and m a c r o _ r e c a l l (see Equation (18)). Thus, the measure m a c r o _ F 1 can be calculated according to Formula (23).
m a c r o _ F 1 = 2 × m a c r o _ p r e c i s i o n × m a c r o _ r e c a l l m a c r o _ p r e c i s i o n + m a c r o _ r e c a l l
In problems with unbalanced classes, where selected classes are more important than others, the m i c r o _ modification leads to the m i c r o _ p r e c i s i o n (see Equation (19)) and m i c r o _ r e c a l l (see Equation (20)). Thus, the measure m i c r o _ F 1 can be calculated according to Formula (24).
m i c r o _ F 1 = 2 × m i c r o _ p r e c i s i o n × m i c r o _ r e c a l l m i c r o _ p r e c i s i o n + m i c r o _ r e c a l l
From Formulas (19), (20) and (24), one can conclude that, for the multi-class classification problem, the values of the measures m i c r o _ p r e c i s i o n , m i c r o _ r e c a l l , and m i c r o _ F 1 are equal to each other. At the same time, the values of these measures are equal to the a c c u r a c y of Formula (16).
The Matthews correlation coefficient (MCC) is only used in classifications of up to two classes (see Equation (10)). For classifications with more than two classes, it is often irrelevant to determine the division of multiple classes into two classes (positive and negative) [45]. Therefore, J. Gorodkin, in his work [46], proposed an extended correlation coefficient (called the R K statistic, for K different classes) that can be used in multi-class classification. Based on this, we defined M C C for multiple classes denoted as Formula (25), where s is the sum of all samples, T P i + F P i is the value of all samples in row i, and T P i + F N i is the value of all samples in column i.
M C C _ m u l t i c l a s s = s × i = 1 c T P i i = 1 c ( ( T P i + F P i ) × ( T P i + F N i ) ) s 2 i = 1 c ( T P i + F P i ) 2 × s 2 i = 1 c ( T P i + F N i ) 2
The last discussed measure is BalancedAccuracy, which could be used to calculate accuracy for unbalanced classes [41]. According to Formula (11), B a l a n c e d A c c u r a c y _ b i n a r y is the arithmetic mean of r e c a l l _ b i n a r y and s p e c i f i c i t y _ b i n a r y . For binary classification, the value of s p e c i f i c i t y _ b i n a r y for the first class equals r e c a l l _ b i n a r y for the second class. For this reason, in multi-class classification, to calculate B a l a n c e d A c c u r a c y , one must count the recall for each class separately and then calculate the arithmetic mean of these values, according to Formula (26).
B a l a n c e d A c c u r a c y = 1 c i = 1 c T P i T P i + F N i

4. Preference-Driven Quality Assessment for Multi-Class Classification

Considering the measures described above, the preferences for individual classes in the classifier quality assessment are insufficient because there is no place in their construction to indicate such preferences. Thus, we propose a new preference-driven classification quality evaluation measure to evaluate classification quality based on different weights for each decision class. The proposed measure works independently of the number of decision classes—its definition is the same in binary and multi-class classification. We also suggest default values for this measure that can be used, in the test case, without specifying exact preferences for each decision class.

4.1. Proposed Preference-Driven Classification Measure

According to the measures given in Section 3.2, the proposed measure was defined based on the confusion matrix given in Table 2. We aim to keep it as simple as possible while satisfying the assumption of adjustment to preferences (of each decision class). Therefore, the proposed measure is based on a confusion matrix and precision and recall measures with κ parameters determining their relative importance.
The preference-driven classification measure is denoted as preference - driven κ , where κ is the preference vector, whose length is equal to the number of decision classes (see Formula (27)). The κ weights for each of the subsequent measures are written on the subsequent positions of the vector. The higher the κ value for a given decision class, the greater the importance of precision (determined by Formula (28)) relative to recall (determined by Formula (29))—based on this class only. Therefore, changing the κ values of a given class makes it possible to control the relative importance of precision and recall. This is a multi-criteria process because the κ value can differ for each class.
Finally, the preference - driven κ measure can be expressed by Formula (30). One should note that κ is a parameter related to the measure by which the relative importance between precision and recall can be established for each decision class separately. For example, κ = [ 0.2 , 0.6 , 0.3 ] means that, for the first class, 20% precision and 80% recall are used; for the second class, 60% precision and 40% recall are used, and for the third class, 30% precision and 70% recall are used. To keep the final value of the preference - driven κ measure in the range [0.0, 1.0], the sum of all these values is divided by the number of classes.
p r e f e r e n c e d r i v e n κ = 1 c i = 1 c κ i × T P i T P i + F P i + ( 1 κ i ) × T P i T P i + F N i
p r e c i s i o n i = T P i T P i + F P i
r e c a l l i = T P i T P i + F N i
p r e f e r e n c e d r i v e n κ = 1 c i = 1 c κ i × p r e c i s i o n i + ( 1 κ i ) × r e c a l l i

4.2. Proposed Measure Analysis in the Test Case

The sample confusion matrix was prepared for a classification problem with two decision classes to demonstrate how the measure works depending on the preference and classification outcome. The two classes were chosen to visualize the measure’s values.
Figure 1 presents nine confusion matrices for which preference - driven κ values were determined for κ , being a combination of all values from 0.0 to 1.0 with a step of 0.1, i.e.,
[ 0.0 , 0.0 ] , [ 0.0 , 0.1 ] , [ 0.5 , 0.4 ] , [ 0.5 , 0.5 ] , [ 0.5 , 0.6 ] , [ 1.0 , 0.9 ] , [ 1.0 , 1.0 ] .
It gives a total of 121 different combinations vectors of preferences. The results are visualized in the following figures.
To better capture the distribution of values for the preference-driven measure, an analysis based on the confusion matrix c m 4 (see Figure 1) was presented. It was selected because the quality assessment in terms of classical measures always means the same values, i.e., precision (see Equation (17)) is 0.7500, and recall (see Equation (18)) is 0.8333. In Figure 2, one can see that the value of the preference-driven measure ranges from 0.5833 for κ = [ 1.0 , 0.0 ] to 1.0000 for κ = [ 0.0 , 1.0 ] . It is possible to obtain exactly the same values for recall ( κ = [ 0.0 , 0.0 ] ) and precision ( κ = [ 1.0 , 1.0 ] ), but depending on the preference, the classifier will be evaluated differently. For example, preference - driven [ 0.1 , 0.3 ] , i.e., for κ 1 = 0.1 and κ 2 = 0.3 is 0.8083 (green dot in Figure 2); similarly, preference - driven [ 0.1 , 0.8 ] = 0.9333 (black dot in Figure 2), preference - driven [ 0.4 , 0.4 ] = 0.7833 (red dot in Figure 2), preference - driven [ 0.9 , 0.1 ] = 0.6250 (light blue dot in Figure 2), and preference - driven [ 0.9 , 0.8 ] = 0.8000 (bright orange dot in Figure 2).
Next, we present different solution spaces obtained for successive confusion matrices. Figure 3 presents such solution spaces for the proposed measure for c m 4 and c m 8 confusion matrices (please refer to Figure 1, which can be described as an opposition to each other, and thus they were selected for the analysis. The value of κ 1 increases, and the value of κ 2 decreases (the value of the proposed measure for c m 4 decreases, while that for c m 8 increases—and vice versa).
Different solution spaces for the confusion matrices from Figure 1 are presented in Figure 4 and Figure 5. In the first case (Figure 4), specific confusion matrices are selected:
  • c m 1 , which represents the ideal classification—in Figure 4a, it can be seen that the value of the preference - driven κ measure is the same for each κ vector;
  • c m 9 , which represents a classification in which all samples are assigned to only one class (in this case, the first one)—in Figure 4b, it can be seen that κ 2 does not affect the value of the preference - driven κ measure;
  • c m 4 and c m 8 , which were previously presented in Figure 3, but this time are shown separately in Figure 4c,d, respectively—they represent an example, not an extreme, case of classification.
Figure 5 presents the solution spaces for all (except c m 9 ) confusion matrices from Figure 1. It allowed us to present the dynamics of the solution space based on the preference - driven κ , depending on the classification being evaluated and the value of the κ vector. Depending on the values of the κ vector, the proposed measure differently evaluates the classifier. It is also presented in Table 3, where the results for different values of the κ vector are presented. Examples [ 0.3 , 0.6 ] , [ 0.9 , 0.4 ] , and [ 0.5 , 0.5 ] have been chosen, as well as precision (which is equivalent to κ = [ 0.0 , 0.0 ] ) and recall (equivalent to κ = [ 1.0 , 1.0 ] ). As can be seen, depending on the given preferences, the evaluation of the classifier even in this case changes noticeably.
Analogously to Table 3, Figure 6 presents the classification quality assessment values for each of the prepared confusion matrices. Such a visualization allows us to observe the influence of the proposed measure on the classification evaluation (compare with the earlier discussed example for c m 4 and c m 8 for which precision, recall, and F1 have identical values), where diagonal lines allow a more straightforward analysis of changes between sample confusion matrices; in this case, we can see that the mentioned measures have identical values for these confusion matrices, but the proposed preference-driven measure obtains different scores for the same preference vectors. Please note that confusion matrices should not be interpreted as consecutive occurrences. Classification evaluation values should be compared regardless of the order in which they are presented.
To further compare the proposed preference-driven measure with the F1 measure, which is also based on recall and precision values, we analyzed the values of the preference vector ( κ ) that produce the same classification score as the F1.
We performed careful analyses for all the confusion matrices described in this section, except for c m 1 , for which the value of all measures is always 1.0000 (this confusion matrix represents the ideal classification). Our observations indicate that there are (unlike the recall and precision measures) no classical values for the preference vector to find the equivalent value of the F1 measure. Therefore, for subsequent confusion matrices, the F1 counterparts are, respectively, preference-driven measures with preference vectors: for c m 2 , it is κ = [ 0.015 , 0.095 ] , then for c m 3 , it is κ = [ 0.010 , 0.145 ] ; for c m 4 , it is κ = [ 0.189 , 0.284 ] ; for c m 5 , it is κ = [ 0.045 , 0.095 ] ; for c m 6 , it is κ = [ 0.06 , 0.12 ] ; for c m 7 , it is κ = [ 0.5 , 0.5 ] ; for c m 8 , it is κ = [ 0.284 , 0.189 ] , and for c m 9 , it is κ = [ 0.67 , 0.00 ] .
These observations indicate significant differences between the F1 and the proposed preference-driven measures. As we have already presented, F1 for a single confusion matrix (i.e., the classifier score) is always represented by a single value. In contrast, it is possible to control the quality score according to preferences in the preference-driven case.

4.3. Default Values of the Preference Vector

The proposed measure is used to evaluate the classifier’s quality as closely as possible to the stated preferences (as long as the decision maker clearly describes these preferences). To make the measure more comparable with other measures, we propose, as an alternative, the default values of the preference vector for the preference - driven κ measure.
The ratio of the number of objects in each class to the number of all samples is taken as the default value of the preference vector. This means that, for classes with many samples, higher weight is related to precision for this class, while, for classes with a relatively small number of samples, higher weight is related to recall of this class. Such a solution allows compensation for the situation in which the samples are more often classified into classes with many samples. When validating a classifier, there is always a learning set (whether for train-and-test or cross-validation), which each time allows the mentioned default values of the preference vector to be determined—the preference vector can correspond to the distribution of cases in the learning data.
In analogy with previously adopted designations, a notation for the default value of the κ vector in Formula (31) is proposed. The designations are the same as those contained in Table 2; additionally, s is the sum of all samples.
κ = T P 1 + F N 1 s , T P 2 + F N 2 s , , T P i + F N i s , , T P C 1 + F N C 1 s , T P C + F N C s
For example, in the situation described in Section 4.2, where the classes in the confusion matrix are at equilibrium, the implicit vector would be of the form κ = [ 0.5 , 0.5 ] . Its value is incidentally given in Table 3, in column preference - driven [ 0.5 , 0.5 ] .
Extending the example, a confusion matrix is proposed, shown in Table 4. In this case, there are three classes, containing 50, 20, and 30 samples each, respectively. Therefore, the default values of the preference vector would be as follows:
κ = 50 100 , 20 100 , 30 100 ,
so
κ = [ 0.5 , 0.2 , 0.3 ] .
Here, 50 out of 100 instances are in class 1, 20 out of 100 instances are in class 2, and 30 out of 100 instances are in class 3. The preference-driven measure value is 0.656, because
preference - driven [ 0.5 , 0.2 , 0.3 ] = 1 3 × 0.5 × T P 1 T P 1 + F P 1 + ( 1 0.5 ) × T P 1 T P 1 + F N 1 + 1 3 × 0.2 × T P 2 T P 2 + F P 2 + ( 1 0.2 ) × T P 2 T P 2 + F N 2 + 1 3 × 0.3 × T P 3 T P 3 + F P 3 + ( 1 0.3 ) × T P 3 T P 3 + F N 3
so
preference - driven [ 0.5 , 0.2 , 0.3 ] = 1 3 × 0.5 × 40 40 + 8 + 9 + 0.5 × 40 40 + 7 + 3 + 1 3 × 0.2 × 10 10 + 7 + 1 + 0.8 × 10 10 + 8 + 2 + 1 3 × 0.3 × 20 20 + 3 + 2 + 0.7 × 20 20 + 9 + 1
preference - driven [ 0.5 , 0.2 , 0.3 ] = 1 3 × ( 0.5 × 0.702 + 0.5 × 0.8 + 0.2 × 0.556 + 0.8 × 0.5 + 0.3 × 0.8 + 0.7 × 0.667 )
preference - driven [ 0.5 , 0.2 , 0.3 ] = 1 3 × ( 0.351 + 0.4 + 0.111 + 0.4 + 0.24 + 0.467 )
preference - driven [ 0.5 , 0.2 , 0.3 ] = 0.656
The results for the default values of the preference vector for the preference - driven κ measure are also analyzed in Section 5, when testing the proposed measure with different real data sets.

5. Analysis on Real-World Data Sets

As the paper proposes a new classification quality assessment measure whose value depends on the stated preferences (called preference vector), we conducted experiments on real-world data sets. First, we checked the importance of the proposed measure depending on the given preference vector—while comparing the performance and ranks of the classifiers with classical measures of classification quality assessment.

5.1. Experiment Conditions

The proposed measure preference - driven κ is compared with the classical measures described in this paper (see Section 3.2). As some of the measures, in the case of multi-class classification, reduce to the same measure, the following names are used in this section: accuracy (see Equation (16)), precision (see Equation (17)), recall (see Equation (18)), F1 (see Equation (23)), and MCC (see Equation (25)).
Four well-known data sets were selected for classification, whose structures are described in Table 5. As one can see, multi-class data sets were selected, in which, additionally, the distribution of samples in classes was uneven (see “percent of samples per class” in Table 5, where the ratio of samples of each class to the whole data set is given). On the other hand, as test classifiers, we selected classifiers available in the system Weka-3-6-11 [47]. More specifically, these were the following classifiers: Bagging [48], BayesNet [49], DecisionTable [50], C4.5 (J48) [51], and RandomForest [52]. In each case, 10-fold cross-validation was used so that each sample was subject to prediction.
Classification results from the Weka system (including confusion matrices) and the values of the proposed classification quality evaluation measure for fixed preference vectors are available on the website of the Department of Machine Learning of the University of Economics in Katowice (https://www.ue.katowice.pl/jednostki/katedry/wiik/katedra-uczenia-maszynowego/zrodla-danych/preference-driven.html (accessed on 6 February 2022)).
The number of decision classes in each data set is different, so the checked combinations of values in the preference vector determined different, possible values of subsequent elements of the vector. Thus, the number of combinations is close to 15,000 (except for the krkopt data set, which is too large and had to exceed this value). In this way, we obtained:
  • car—[0, 0.1, 0.2, 0.3, 0.4, 0.5, 0.6, 0.7, 0.8, 0.9, 1], which gives 14,641 combinations (because: 4 decision classes and 11 values of κ , so 11 4 = 14,641);
  • nursery—[0, 0.166, 0.332, 0.498, 0.664, 0.83, 1], which gives 7 5 = 16,807 combinations;
  • dermatology—[0, 0.25, 0.5, 0.75, 1], which gives 5 6 = 15,625; combinations;
  • krkopt—[0.33, 0.66], which gives 2 18 = 262,144 combinations.

5.2. Experimental Results

Table 6, Table 7, Table 8 and Table 9 present the results for all data sets and all classifiers. Ratings have been made for all measures, with preference - driven κ determined each time for the default values and five different, selected preference vectors. Evaluation values were presented along with the ranking order of the classifier for each measure (in brackets). This approach allows us not only to evaluate the differences between the evaluation values but also to indicate whether, using the proposed measure, it could happen that another classifier would be better than those indicated by the classical classification quality evaluation measures. On the other hand, Figure 7 shows histograms of the preference - driven κ measure values for all tested preference vectors.
Figure 7 presents 20 different histograms used to show the distribution of values (without any unnecessarily detailed information). In part (a), successive histograms are shown for the car data set, then (b) contains the nursery histograms, (c) is related to dermatology, and (d) is the krkopt results. This figure concerns the value of the preference-driven measure determined for different preference vectors ( κ ). Therefore, the X-axis represents the value of the preference-driven measure, while the Y-axis represents the number of occurrences of this value.
This presentation in Figure 7 allows us to notice that in each case, using the proposed measure, it is possible to obtain a different score with the same confusion matrix. In addition, in most cases, the values of the measure are close to a normal distribution. A slightly more interesting case is (b), with the Bagging, BayesNet, and RandomForest classifiers. In this case, however, it should be noted that the classification each time gives a specific confusion matrix (detailed results are available on the UE Katowice website (https://www.ue.katowice.pl/jednostki/katedry/wiik/katedra-uczenia-maszynowego/zrodla-danych/preference-driven.html) (accessed on 6 February 2022)). Note that there are only two samples in the first class. The class with the most significant number of samples is always correctly identified. Additionally, in case (c), the particular histogram is obtained for the RandomForest classifier, where the classification was excellent, close to error-free. In contrast, the values were rounded to the second decimal for the histogram, hence such a high concentration of preference-driven measure values.
Analyzing the experimental results, it is worth noting that by using different vectors, in the case of a preference-driven measure, it is possible to indicate a different classifier as more adapted to the problem. For example, in the case of the car data set (see Table 6), the classifier DecisionTable turns out to be the best for some preference vectors (e.g., [ 0 , 1 , 1 , 0.9 ] ); similarly, the classifier BayesNet turns out to be better than Bagging if the preference vector is, among others, [ 0.3 , 1 , 1 , 1 ] . The situation is similar for the set of nursery data (see Table 7), where also DecisionTable turns out to be the best, while the preference vector is [ 0.498 , 1 , 0.166 , 1 , 0 ] . A similar difference, but on subsequent ranks (between 3 and 4), can be observed also in the case of the krkopt data set (see Table 9). Although, in the case of the dermatology data set (see Table 8), no change of ranks was observed, it could be observed that the disproportion between the assessment of specific classifiers changed a lot and, e.g., in the case of vector [ 0 , 0 , 0 , 1 , 1 , 0.75 ] , the difference between the best BayesNet and the second RandomForest was less than 0.0083, where, with classical measures of classification quality assessment, this difference was around 0.02.
The aim of the experiment was to check whether the proposed measure will show different classifiers for the same problem depending on the value in the preference vector. It turned out that, indeed, the measure indicates different classifiers, even with a limited number of vectors tested.

6. Discussion

In supervised learning, one of the necessary steps in adequately designed research is teaching on training data different classifiers, sometimes with different parameters, and then evaluating their quality. In our work, we assume that the user has specific preferences. In addition, they relate to particular decision classes. Thus, we can immediately identify a suitable classifier.
Using a vector of preferences, it is possible to indicate whether precision or recall is more important. It is crucial for many problems to indicate that it is possible to select a classifier better adapted to the structure of the data and decision makers’ expectations for each class separately. The best strategy is to examine the training of as many classifiers as possible (with different parameters). Such an approach allows the identification of the potentially best classifiers for a given problem.
The proposed preference-driven measure used in the experiments allowed for a better selection of the classifier most suitable for the task. With an approach allowing us to indicate whether precision or recall is more important—for each class separately—it is possible to select a classifier more adapted to the more important (from the evaluation point of view) decision class.
Despite the limited number of combinations, the conducted experiments indicated that the proposed measure, depending on the preferences conveyed in the form of a preference vector, points to different classifiers as the best choice for further prediction.
Since the set of values in the preference vector is infinite, the measurement values are also unlimited. Therefore, the calculation of a measure and comparison with other measures is possible only by calculating them at specific points. It should be noted that it is not necessary to test many combinations of the preference vector in a real application. Instead, these should be predetermined, indicating the relative importance of the recall–precision balance in each decision class.
This also distinguishes the preference-driven measure from the F β measure, which also raised attempts to weigh between precision and recall. However, the proposed preference-driven measure is different. Note that, in the case of F β , there is a weighting between precision and recall overall for the classifier. In contrast, there are weights for each decision class with the preference-driven measure. It allows us to change the emphasis on precision and recall depending on each decision class (differently). In the case of F β , this possibility does not exist.

7. Conclusions

This article presents a new idea for a preference-driven classification measure. We tried to show that the measure works, i.e.,
  • for different preference vectors, different classifiers are more advantageous then others;
  • the obtained results of the comparison make it easier for the user to understand the effects of classification and make the right decision as to which classifier to use.
In the “objective” approach (without preferences), the result is unambiguous and comparable, but the best classifier will not necessarily be adjusted to the subjective needs of the user. In the “subjective” approach (preference-driven), comparability is difficult, but in return, users can acquire a classifier better suited to their requirements.
The concept of this measure results from the shortcomings of measures related to the multi-class classification. Nowadays, we observe a large number of classification methods. There is no single versatile classification measure capable of catching up on concepts related to both: overall good classification quality and a particular focus on the selected decision classes. The whole idea of different classification measures is mostly extended for the binary decision classes, which often fail to achieve good results for real-world data. At the same time, multi-class classification measures are based on averaging the results, which can be fair for general cases but fails to include decision makers’ preferences related to the particular classes.
Similarly, for unbalanced cases, where there is a need to focus on particular classes, the proposed preference-driven measure fits well for this gap. To be more precise, our proposed preference-driven measure can be aligned with the decision makers’ preferences regarding the relative importance of precision and recall. We also present the idea of setting the default values for the vector of preferences based on the overall number of samples assigned to every decision class. The most important advantage of the proposed idea is the good fit between the well-known measures such as precision and recall. Moreover, the κ preference vector allows us to direct the focus to a particular decision class, or even to express the importance of selected decision classes in terms of the precision measure (and others in terms of recall).
At the same time, we show that even for potentially trivial cases, such preference-driven measures could lead to entirely different results based on the κ selection. It opens a discussion for multi-class classification and leads to an interesting situation. The solution for the classification problem should not be considered a single scalar value.
In the future, a preference-driven measure can be used in line with the proposed approach. Alternatively, the factors of which the measure is composed could be scrutinized, and other measures could be used instead of the class’s relative precision and recall values.

Author Contributions

Conceptualization, J.K.; methodology, J.K., B.P., K.K. and P.J.; software, J.K.; validation, J.K. and B.P.; formal analysis, J.K., B.P., K.K. and P.J.; investigation, J.K. and B.P.; resources, J.K.; data curation, J.K.; writing—original draft preparation, J.K., B.P., K.K. and P.J.; writing—review and editing, J.K., B.P., K.K. and P.J.; visualization, J.K.; supervision, J.K.; project administration, J.K., B.P. and P.J. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The classification results (including confusion matrices) obtained in the Weka program, as well as the values of the preference-driven measure for the analyzed preference vectors, can be found on the website of the Department of Machine Learning of the University of Economics in Katowice: https://www.ue.katowice.pl/jednostki/katedry/wiik/katedra-uczenia-maszynowego/zrodla-danych/preference-driven.html (accessed on 6 February 2022). The implementation of the proposed preference-driven measure was prepared in Python language. The source code of this implementation is publicly available via GitHub: https://github.com/jankozak/preference-driven (accessed on 6 February 2022).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Gösgens, M.; Zhiyanov, A.; Tikhonov, A.; Prokhorenkova, L. Good Classification Measures and How to Find Them. In Proceedings of the 35th Conference on Neural Information Processing Systems (NeurIPS 2021), Online, 6–14 December 2021. [Google Scholar]
  2. Seijo-Pardo, B.; Porto-Díaz, I.; Bolón-Canedo, V.; Alonso-Betanzos, A. Ensemble feature selection: Homogeneous and heterogeneous approaches. Knowl.-Based Syst. 2017, 118, 124–139. [Google Scholar] [CrossRef]
  3. Lewis, D.D.; Catlett, J. Heterogeneous uncertainty sampling for supervised learning. In Machine Learning Proceedings 1994; Elsevier: Amsterdam, The Netherlands, 1994; pp. 148–156. [Google Scholar]
  4. Campagner, A.; Sconfienza, L.; Cabitza, F. H-accuracy, an alternative metric to assess classification models in medicine. In Digital Personalized Health and Medicine; IOS Press: Amsterdam, The Netherlands, 2020; pp. 242–246. [Google Scholar]
  5. Gilli, M.; Schumann, E. Accuracy and precision in finance. Available SSRN 2698114 2015. [Google Scholar] [CrossRef]
  6. Canbek, G.; Taskaya Temizel, T.; Sagiroglu, S. BenchMetrics: A systematic benchmarking method for binary classification performance metrics. Neural Comput. Appl. 2021, 33, 14623–14650. [Google Scholar] [CrossRef]
  7. Amershi, S.; Cakmak, M.; Knox, W.B.; Kulesza, T. Power to the people: The role of humans in interactive machine learning. Ai Mag. 2014, 35, 105–120. [Google Scholar] [CrossRef] [Green Version]
  8. Wu, X.; Xiao, L.; Sun, Y.; Zhang, J.; Ma, T.; He, L. A Survey of Human-in-the-loop for Machine Learning. arXiv 2021, arXiv:2108.00941. [Google Scholar]
  9. Talbot, J.; Lee, B.; Kapoor, A.; Tan, D.S. EnsembleMatrix: Interactive visualization to support machine learning with multiple classifiers. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems, Boston, MA, USA, 4–9 April 2009; pp. 1283–1292. [Google Scholar]
  10. Green, B.; Chen, Y. The principles and limits of algorithm-in-the-loop decision making. Proc. ACM Hum. -Comput. Interact. 2019, 3, 1–24. [Google Scholar] [CrossRef] [Green Version]
  11. Kononenko, I.; Bratko, I. Information-Based Evaluation Criterion for Classifier’s Performance. Mach. Learn. 1991, 6, 67–80. [Google Scholar] [CrossRef]
  12. Valverde-Albacete, F.J.; Peláez-Moreno, C. 100% classification accuracy considered harmful: The normalized information transfer factor explains the accuracy paradox. PLoS ONE 2014, 9, e84217. [Google Scholar] [CrossRef] [Green Version]
  13. Saito, T.; Rehmsmeier, M. The precision-recall plot is more informative than the ROC plot when evaluating binary classifiers on imbalanced datasets. PLoS ONE 2015, 10, e0118432. [Google Scholar] [CrossRef] [Green Version]
  14. Ben-David, A. A lot of randomness is hiding in accuracy. Eng. Appl. Artif. Intell. 2007, 20, 875–885. [Google Scholar] [CrossRef]
  15. Sokolova, M.; Japkowicz, N.; Szpakowicz, S. Beyond accuracy, F-score and ROC: A family of discriminant measures for performance evaluation. In Australasian Joint Conference on Artificial Intelligence; AAAI Workshop-Technical Report; Springer: Berlin/Heidelberg, Germany, 2006; Volume 4304, pp. 24–29. [Google Scholar] [CrossRef] [Green Version]
  16. Grandini, M.; Bagli, E.; Visani, G. Metrics for multi-class classification: An overview. arXiv 2020, arXiv:2008.05756. [Google Scholar]
  17. Hossin, M.; Sulaiman, M. A Review on Evaluation Metrics for Data Classification Evaluations. Int. J. Data Min. I Knowl. Manag. Process 2015, 5, 1–11. [Google Scholar] [CrossRef]
  18. Ferri, C.; Hernández-Orallo, J.; Modroiu, R. An experimental comparison of performance measures for classification. Pattern Recognit. Lett. 2009, 30, 27–38. [Google Scholar] [CrossRef]
  19. Emmert-Streib, F.; Moutari, S.; Dehmer, M. A comprehensive survey of error measures for evaluating binary decision making in data science. Wiley Interdiscip. Rev. Data Min. Knowl. Discov. 2019, 9, 1–15. [Google Scholar] [CrossRef] [Green Version]
  20. Sokolova, M.; Lapalme, G. A systematic analysis of performance measures for classification tasks. Inf. Process. Manag. 2009, 45, 427–437. [Google Scholar] [CrossRef]
  21. Demšar, J. Statistical comparisons of classifiers over multiple data sets. J. Mach. Learn. Res. 2006, 7, 1–30. [Google Scholar]
  22. Tharwat, A. Classification assessment methods. Appl. Comput. Inform. 2018, 17, 168–192. [Google Scholar] [CrossRef]
  23. Zhang, X.; Li, X.; Feng, Y. A classification performance measure considering the degree of classification difficulty. Neurocomputing 2016, 193, 81–91. [Google Scholar] [CrossRef]
  24. Yu, S.; Li, X.; Feng, Y.; Zhang, X.; Chen, S. An instance-oriented performance measure for classification. Inf. Sci. 2021, 580, 598–619. [Google Scholar] [CrossRef]
  25. Gong, M. A Novel Performance Measure for Machine Learning Classification. Int. J. Manag. Inf. Technol. 2021, 13, 11–19. [Google Scholar] [CrossRef]
  26. Carbonero-Ruz, M.; Martínez-Estudillo, F.J.; Fernández-Navarro, F.; Becerra-Alonso, D.; Martínez-Estudillo, A.C. A two dimensional accuracy-based measure for classification performance. Inf. Sci. 2017, 382–383, 60–80. [Google Scholar] [CrossRef]
  27. Kasperczuk, A.; Dardzinska, A. Automatic system for IBD diagnosis. Procedia Comput. Sci. 2021, 192, 2863–2870. [Google Scholar] [CrossRef]
  28. Bac, C.W.; Hemming, J.; Van Henten, E.J. Robust pixel-based classification of obstacles for robotic harvesting of sweet-pepper. Comput. Electron. Agric. 2013, 96, 148–162. [Google Scholar] [CrossRef]
  29. Meng, J.; Zhang, W. Volume measure in 2DPCA-based face recognition. Pattern Recognit. Lett. 2007, 28, 1203–1208. [Google Scholar] [CrossRef]
  30. Burduk, R. Classification Performance Metric for Imbalance Data Based on Recall and Selectivity Normalized in Class Labels. arXiv 2020, arXiv:2006.13319. [Google Scholar]
  31. Hand, D.J.; Christen, P.; Kirielle, N. F*: An interpretable transformation of the F-measure. Mach. Learn. 2021, 110, 451–456. [Google Scholar] [CrossRef] [PubMed]
  32. Mitchell, T.M. Machine Learning, International Edition; McGraw-Hill Series in Computer Science; McGraw-Hill Education: New York, NY, USA, 1997. [Google Scholar]
  33. Townsend, J.T. Theoretical analysis of an alphabetic confusion matrix. Percept. Psychophys. 1971, 9, 40–50. [Google Scholar] [CrossRef]
  34. Provost, F.; Kohavi, R. Glossary of terms. J. Mach. Learn. 1998, 30, 271–274. [Google Scholar] [CrossRef]
  35. Room, C. Confusion Matrix. Mach. Learn. 2019, 6, 27. [Google Scholar]
  36. Lee, N.; Yang, H.; Yoo, H. A surrogate loss function for optimization of Fβ score in binary classification with imbalanced data. arXiv 2021, arXiv:2104.01459. [Google Scholar]
  37. Van Rijsbergen, C.J. Information Retrieval; Butterworth-Heinemann: Newton, MA, USA, 1979. [Google Scholar]
  38. Buckland, M.; Gey, F. The relationship between recall and precision. J. Am. Soc. Inf. Sci. 1994, 45, 12–19. [Google Scholar] [CrossRef]
  39. Chicco, D.; Jurman, G. The advantages of the Matthews correlation coefficient (MCC) over F1 score and accuracy in binary classification evaluation. BMC Genom. 2020, 21, 1–13. [Google Scholar] [CrossRef] [Green Version]
  40. Matthews, B.W. Comparison of the predicted and observed secondary structure of T4 phage lysozyme. Biochim. Biophys. Acta (BBA)-Protein Struct. 1975, 405, 442–451. [Google Scholar] [CrossRef]
  41. Brodersen, K.H.; Ong, C.S.; Stephan, K.E.; Buhmann, J.M. The Balanced Accuracy and Its Posterior Distribution. In Proceedings of the 2010 20th International Conference on Pattern Recognition, Istanbul, Turkey, 23–26 August 2010; pp. 3121–3124. [Google Scholar] [CrossRef]
  42. Parikh, R.; Mathai, A.; Parikh, S.; Sekhar, G.C.; Thomas, R. Understanding and using sensitivity, specificity and predictive values. Indian J. Ophthalmol. 2008, 56, 45–50. [Google Scholar] [CrossRef] [PubMed]
  43. Tsoumakas, G.; Katakis, I. Multi-label classification: An overview. Int. J. Data Warehous. Min. (IJDWM) 2007, 3, 1–13. [Google Scholar] [CrossRef] [Green Version]
  44. Takahashi, K.; Yamamoto, K.; Kuchiba, A.; Koyama, T. Confidence interval for micro-averaged F1 and macro-averaged F1 scores. Appl. Intell. 2022, 28, 4961–4972. [Google Scholar] [CrossRef] [PubMed]
  45. Jurman, G.; Riccadonna, S.; Furlanello, C. A comparison of MCC and CEN error measures in multi-class prediction. PLoS ONE 2012, 7, e41882. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  46. Gorodkin, J. Comparing two K-category assignments by a K-category correlation coefficient. Comput. Biol. Chem. 2004, 28, 367–374. [Google Scholar] [CrossRef]
  47. Hall, M.; Frank, E.; Holmes, G.; Pfahringer, B.; Reutemann, P.; Witten, I.H. The WEKA data mining software: An update. ACM SIGKDD Explor. Newsl. 2009, 11, 10–18. [Google Scholar] [CrossRef]
  48. Breiman, L. Bagging predictors. Mach. Learn. 1996, 24, 123–140. [Google Scholar] [CrossRef] [Green Version]
  49. Bouckaert, R.R. Bayesian Network Classifiers in Weka; Working Paper No. 14/2004; University of Waikato: Hamilton, New Zealand, 2004. [Google Scholar]
  50. Kohavi, R. The Power of Decision Tables. In Proceedings of the 8th European Conference on Machine Learning, Crete, Greece, 25–27 April 1995; Springer: Berlin/Heidelberg, Germany, 1995; pp. 174–189. [Google Scholar]
  51. Quinlan, R. C4.5: Programs for Machine Learning; Morgan Kaufmann Publishers: San Mateo, CA, USA, 1993. [Google Scholar]
  52. Breiman, L. Random Forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Confusion matrix used for measure analysis.
Figure 1. Confusion matrix used for measure analysis.
Entropy 24 00531 g001
Figure 2. All possible values of the preference-driven measure for the confusion matrix c m 4 .
Figure 2. All possible values of the preference-driven measure for the confusion matrix c m 4 .
Entropy 24 00531 g002
Figure 3. All possible values of the preference-driven measure for the confusion matrix c m 4 and its opposite c m 8 .
Figure 3. All possible values of the preference-driven measure for the confusion matrix c m 4 and its opposite c m 8 .
Entropy 24 00531 g003
Figure 4. Example distributions of values of the preference-driven measure: (a) confusion matrix c m 1 ; (b) confusion matrix c m 9 ; (c) confusion matrix c m 4 ; (d) confusion matrix c m 8 .
Figure 4. Example distributions of values of the preference-driven measure: (a) confusion matrix c m 1 ; (b) confusion matrix c m 9 ; (c) confusion matrix c m 4 ; (d) confusion matrix c m 8 .
Entropy 24 00531 g004
Figure 5. Solution spaces of proposed measure values for selected confusion matrices.
Figure 5. Solution spaces of proposed measure values for selected confusion matrices.
Entropy 24 00531 g005
Figure 6. Example values of the proposed measure in comparison with other measures for assessing the quality of classification (the diagonal lines allow easier analysis of the changes between the sample confusion matrices).
Figure 6. Example values of the proposed measure in comparison with other measures for assessing the quality of classification (the diagonal lines allow easier analysis of the changes between the sample confusion matrices).
Entropy 24 00531 g006
Figure 7. Histogram of preference-driven measure values in the (a) car data set; (b) nursery data set; (c) dermatology data set; (d) krkopt data set—the X-axis represents the value of the preference-driven measure, while the Y-axis represents the number of occurrences of this value.
Figure 7. Histogram of preference-driven measure values in the (a) car data set; (b) nursery data set; (c) dermatology data set; (d) krkopt data set—the X-axis represents the value of the preference-driven measure, while the Y-axis represents the number of occurrences of this value.
Entropy 24 00531 g007
Table 1. Confusion matrix for binary classification.
Table 1. Confusion matrix for binary classification.
Predicted Class 1Predicted Class 2
Actual class 1TPFN
Actual class 2FPTN
Table 2. Confusion matrix for multiple classes.
Table 2. Confusion matrix for multiple classes.
Predicted Class 1Predicted Class 2Predicted Class 3Predicted Class iPredicted Class C-1Predicted Class C
F P 2 F P 3 F P i F P C 1 F P C
Actual class 1 T P 1 T N \ { 1 } T N \ { 1 , 2 } T N \ { 1 , 3 } T N \ { 1 , i } T N \ { 1 , C 1 } T N \ { 1 , C }
F N 1 F N 1 F N 1 F N 1 F N 1
F P 1 F P 3 F P i F P C 1 F P C
Actual class 2 T N \ { 1 , 2 } T P 2 T N \ { 2 } T N \ { 2 , 3 } T N \ { 2 , i } T N \ { 2 , C 1 } T N \ { 2 , C }
F N 2 F N 2 F N 2 F N 2 F N 2
F P 1 F P 2 F P i F P C 1 F P C
Actual class 3 T N \ { 1 , 3 } T N \ { 2 , 3 } T P 3 T N \ { 3 } T N \ { 3 , i } T N \ { 3 , C 1 } T N \ { 3 , C }
F N 3 F N 3 F N 3 F N 3 F N 3
F P 1 F P 2 F P 3 F P C 1 F P C
Actual class i T N \ { 1 , i } T N \ { 2 , i } T N \ { 3 , i } T P i T N \ { i } T N \ { i , C 1 } T N \ { i , C }
F N i F N i F N i F N i F N i
F P 1 F P 2 F P 3 F P i F P C
Actual class C-1 T N \ { 1 , C 1 } T N \ { 2 , C 1 } T N \ { 3 , C 1 } T N \ { i , C 1 } T P C 1 T N \ { C 1 } T N \ { C , C 1 }
F N C 1 F N C 1 F N C 1 F N C 1 F N C 1
F P 1 F P 2 F P 3 F P i F P C 1
Actual class C T N \ { 1 , C } T N \ { 2 , C } T N \ { 3 , C } T N \ { i , C } T N \ { C 1 , C } T P C T N \ { C }
F N C F N C F N C F N C F N C
Table 3. Example values of the proposed preference-driven measure in comparison with other measures for assessing the quality of classification (p-d is the abbreviation for the preference-driven measure). Results determined for all confusion matrices presented in Figure 1.
Table 3. Example values of the proposed preference-driven measure in comparison with other measures for assessing the quality of classification (p-d is the abbreviation for the preference-driven measure). Results determined for all confusion matrices presented in Figure 1.
p - d [ 0.3 , 0.6 ] p - d [ 0.9 , 0.4 ] p - d [ 0.5 , 0.5 ] Precision 1Recall 2F1 3
c m 1 41.00001.00001.00001.00001.00001.0000
c m 2 40.93500.86500.90830.91570.90000.9083
c m 3 40.87710.75140.82860.85710.80000.8276
c m 4 40.85000.70000.79170.83330.75000.7895
c m 5 40.72500.67000.70420.70830.70000.7041
c m 6 40.68660.60980.65740.66480.65000.6573
c m 7 40.55850.53660.55030.55050.55000.5502
c m 8 40.72500.90830.79170.83330.75000.7895
c m 9 40.42500.27500.37500.25000.50000.3333
1 m a c r o _ p r e c i s i o n (Equation (17)) is equal to preference - driven [ 1.0 , 1.0 ] . 2 m a c r o _ r e c a l l (Equation (18)) is equal to preference - driven [ 0.0 , 0.0 ] . 3 m a c r o _ F 1 (Equation (23)). 4 See Figure 1.
Table 4. Example confusion matrix used to demonstrate how to determine the proposed preference-driven measure.
Table 4. Example confusion matrix used to demonstrate how to determine the proposed preference-driven measure.
Predicted Class 1Predicted Class 2Predicted Class 3
Actual class 14073
Actual class 28102
Actual class 39120
Table 5. Characteristics of the real data sets used to test the proposed preference-driven measure and comparison with classical measures.
Table 5. Characteristics of the real data sets used to test the proposed preference-driven measure and comparison with classical measures.
Data SetNumber of
Samples
Number of
Attributes
Number of
Classes
Percent of Samples
per Class
car1728640.70 0.22 0.04 0.04
nursery12960850.00 0.33
0.33 0.03 0.31
dermatology3663460.17 0.31 0.20
0.14 0.13 0.05
krkopt280566180.10 0.00 0.00
0.01 0.00 0.01
0.02 0.02 0.02
0.05 0.06 0.07
0.10 0.13 0.15
0.16 0.08 0.01
Table 6. Results for the car data set—the value of the classification quality assessment (in brackets, we give the ranking of the classifier, according to the given measure). The ranking determines the order of the classifier depending on the classification quality rating measure used.
Table 6. Results for the car data set—the value of the classification quality assessment (in brackets, we give the ranking of the classifier, according to the given measure). The ranking determines the order of the classifier depending on the classification quality rating measure used.
BaggingBayesNetDecisionTableC4.5 (J48)RandomForest
accuracy0.9167 (3)0.8571 (5)0.9149 (4)0.9236 (2)0.9462 (1)
precision0.7586 (5)0.7940 (4)0.8557 (2)0.8179 (3)0.8565 (1)
recall0.7665 (4)0.6040 (5)0.8177 (3)0.8289 (2)0.8581 (1)
F10.7625 (4)0.6861 (5)0.8363 (2)0.8233 (3)0.8573 (1)
MCC0.8208 (3)0.6737 (5)0.8101 (4)0.8345 (2)0.8842 (1)
preference-driven 10.7670 (4)0.6031 (5)0.8171 (3)0.8291 (2)0.8584 (1)
p - d [ 0 , 0.8 , 0.6 , 1 ] 0.7757 (4)0.7582 (5)0.8600 (1)0.8295 (3)0.8600 (1)
p - d [ 0 , 1 , 0 , 1 ] 0.8021 (4)0.6853 (5)0.8631 (2)0.8475 (3)0.8647 (1)
p - d [ 0 , 1 , 1 , 0.9 ] 0.7494 (5)0.7954 (4)0.8601 (1)0.8127 (3)0.8499 (2)
p - d [ 0.3 , 1 , 1 , 1 ] 0.7555 (5)0.8012 (4)0.8609 (1)0.8160 (3)0.8534 (2)
p - d [ 0.6 , 1 , 0.8 , 1 ] 0.7664 (5)0.7743 (4)0.8587 (1)0.8233 (3)0.8572 (2)
1 Default value of the preference vector (see Section 4.3).
Table 7. Results for the nursery data set—the value of the classification quality assessment (in brackets, we give the ranking of the classifier, according to the given measure). The ranking determines the order of the classifier depending on the classification quality rating measure used.
Table 7. Results for the nursery data set—the value of the classification quality assessment (in brackets, we give the ranking of the classifier, according to the given measure). The ranking determines the order of the classifier depending on the classification quality rating measure used.
BaggingBayesNetDecisionTableC4.5 (J48)RandomForest
accuracy0.9737 (2)0.9033 (5)0.9470 (4)0.9705 (3)0.9909 (1)
precision0.7518 (3)0.7250 (5)0.7661 (2)0.7453 (4)0.7849 (1)
recall0.7226 (3)0.5666 (5)0.6722 (4)0.7313 (2)0.7765 (1)
F10.7369 (3)0.6361 (5)0.7160 (4)0.7382 (2)0.7806 (1)
MCC0.9614 (2)0.8579 (5)0.9234 (4)0.9568 (3)0.9867 (1)
preference-driven 10.7224 (3)0.5671 (5)0.6727 (4)0.7312 (2)0.7764 (1)
p - d [ 0 , 0 , 0 , 1 , 0 ] 0.7550 (3)0.7362 (5)0.7733 (2)0.7470 (4)0.7857 (1)
p - d [ 0 , 0 , 0 , 1 , 1 ] 0.7524 (2)0.7412 (5)0.7520 (3)0.7428 (4)0.7850 (1)
p - d [ 0 , 0.83 , 0 , 1 , 0 ] 0.7545 (3)0.7228 (5)0.7850 (2)0.7490 (4)0.7856 (1)
p - d [ 0.498 , 1 , 0.166 , 1 , 0 ] 0.7544 (3)0.7200 (5)0.7874 (1)0.7495 (4)0.7855 (2)
p - d [ 0.83 , 0.332 , 0.166 , 0 , 1 ] 0.7198 (3)0.5662 (5)0.6555 (4)0.7280 (2)0.7757 (1)
1 Default value of the preference vector (see Section 4.3).
Table 8. Results for the dermatology data set—the value of the classification quality assessment (in brackets, we give the ranking of the classifier, according to the given measure). The ranking determines the order of the classifier depending on the classification quality rating measure used.
Table 8. Results for the dermatology data set—the value of the classification quality assessment (in brackets, we give the ranking of the classifier, according to the given measure). The ranking determines the order of the classifier depending on the classification quality rating measure used.
BaggingBayesNetDecisionTableC4.5 (J48)RandomForest
accuracy0.3798 (5)0.9727 (1)0.8388 (4)0.9454 (3)0.9536 (2)
precision0.3798 (5)0.9692 (1)0.8549 (4)0.9372 (3)0.9492 (2)
recall0.3947 (5)0.9707 (1)0.8204 (4)0.9368 (3)0.9477 (2)
F10.3871 (5)0.9699 (1)0.8373 (4)0.9370 (3)0.9484 (2)
MCC0.2170 (5)0.9660 (1)0.7982 (4)0.9316 (3)0.9418 (2)
preference-driven 10.3923 (5)0.9708 (1)0.8217 (4)0.9368 (3)0.9477 (2)
p - d [ 0 , 0 , 0 , 0 , 1 , 0 ] 0.4090 (5)0.9589 (1)0.8030 (4)0.9400 (3)0.9506 (2)
p - d [ 0 , 0 , 0 , 1 , 1 , 0.75 ] 0.4014 (5)0.9589 (1)0.8430 (4)0.9369 (3)0.9506 (2)
p - d [ 0 , 0 , 1 , 1 , 0 , 1 ] 0.3807 (5)0.9707 (1)0.8800 (4)0.9315 (3)0.9477 (2)
p - d [ 0.5 , 0 , 0.25 , 0.5 , 0.5 , 1 ] 0.3823 (5)0.9699 (1)0.8575 (4)0.9375 (3)0.9492 (2)
p - d [ 0.5 , 1 , 0.75 , 0.5 , 0.5 , 0 ] 0.3921 (5)0.9699 (1)0.8178 (4)0.9364 (3)0.9477 (2)
1 Default value of the preference vector (see Section 4.3).
Table 9. Results for the krkopt data set—the value of the classification quality assessment (in brackets, we give the ranking of the classifier, according to the given measure). The ranking determines the order of the classifier depending on the classification quality rating measure used.
Table 9. Results for the krkopt data set—the value of the classification quality assessment (in brackets, we give the ranking of the classifier, according to the given measure). The ranking determines the order of the classifier depending on the classification quality rating measure used.
BaggingBayesNetDecisionTableC4.5 (J48)RandomForest
accuracy0.5872 (2)0.3607 (5)0.4908 (4)0.5658 (3)0.7025 (1)
precision0.5735 (3)0.3579 (5)0.5784 (2)0.5547 (4)0.7377 (1)
recall0.5406 (2)0.2982 (5)0.5187 (3)0.5178 (4)0.6628 (1)
F10.5566 (2)0.3253 (5)0.5469 (3)0.5356 (4)0.6982 (1)
MCC0.5377 (2)0.2784 (5)0.4300 (4)0.5135 (3)0.6669 (1)
preference-driven 10.5404 (2)0.2967 (5)0.5191 (3)0.5177 (4)0.6629 (1)
preference - driven κ 20.5776 (3)0.3717 (5)0.5806 (2)0.5576 (3)0.7426 (1)
preference - driven κ  30.5542 (2)0.3176 (5)0.5415 (3)0.5288 (4)0.6869 (1)
preference - driven κ  40.5554 (2)0.3262 (5)0.5462 (3)0.5396 (4)0.6925 (1)
preference - driven κ  50.5741 (2)0.3467 (5)0.5472 (4)0.5485 (3)0.7287 (1)
preference - driven κ  60.5638 (2)0.3461 (5)0.5587 (3)0.5410 (4)0.7033 (1)
1 Default value of the preference vector (see Section 4.3). 2 κ = [0, 1, 1, 1, 1, 1, 0.9, 0.9, 1, 0.9, 0.9, 0.9, 0, 0, 0, 0, 0.9, 1]. 3 κ = [0.33, 0.66, 0.33, 0.33, 0.66, 0.33, 0.66, 0.66, 0.33, 0.66, 0.66, 0.33, 0.66, 0.66, 0.66, 0.66, 0.33, 0.33]. 4 κ = [0.66, 0.33, 0.33, 0.33, 0.33, 0.66, 0.66, 0.33, 0.66, 0.66, 0.66, 0.33, 0.33, 0.66, 0.33, 0.33, 0.33, 0.33]. 5 κ = [1, 0.1, 1, 1, 1, 1, 0.1, 0.1, 0, 0.1, 0.1, 0.1, 0, 0, 0, 0, 0.9, 1]. 6 κ = [1, 1, 0.9, 0.9, 0.8, 0.8, 0.7, 0.6, 0.5, 0.4, 0.3, 0.3, 0.2, 0.2, 0.1, 0.1, 0.0, 0.0].
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Kozak, J.; Probierz, B.; Kania, K.; Juszczuk, P. Preference-Driven Classification Measure. Entropy 2022, 24, 531. https://0-doi-org.brum.beds.ac.uk/10.3390/e24040531

AMA Style

Kozak J, Probierz B, Kania K, Juszczuk P. Preference-Driven Classification Measure. Entropy. 2022; 24(4):531. https://0-doi-org.brum.beds.ac.uk/10.3390/e24040531

Chicago/Turabian Style

Kozak, Jan, Barbara Probierz, Krzysztof Kania, and Przemysław Juszczuk. 2022. "Preference-Driven Classification Measure" Entropy 24, no. 4: 531. https://0-doi-org.brum.beds.ac.uk/10.3390/e24040531

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop