Next Article in Journal
Experimental Investigation and Fault Diagnosis for Buckled Wet Clutch Based on Multi-Speed Hilbert Spectrum Entropy
Previous Article in Journal
Continuous Viewpoint Planning in Conjunction with Dynamic Exploration for Active Object Recognition
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Exact Learning Augmented Naive Bayes Classifier

Graduate School of Informatics and Engineering, The University of Electro-Communications, 1-5-1, Chofugaoka, Chofu-shi, Tokyo 182-8585, Japan
*
Author to whom correspondence should be addressed.
Submission received: 29 November 2021 / Revised: 16 December 2021 / Accepted: 17 December 2021 / Published: 20 December 2021
(This article belongs to the Topic Machine and Deep Learning)

Abstract

:
Earlier studies have shown that classification accuracies of Bayesian networks (BNs) obtained by maximizing the conditional log likelihood (CLL) of a class variable, given the feature variables, were higher than those obtained by maximizing the marginal likelihood (ML). However, differences between the performances of the two scores in the earlier studies may be attributed to the fact that they used approximate learning algorithms, not exact ones. This paper compares the classification accuracies of BNs with approximate learning using CLL to those with exact learning using ML. The results demonstrate that the classification accuracies of BNs obtained by maximizing the ML are higher than those obtained by maximizing the CLL for large data. However, the results also demonstrate that the classification accuracies of exact learning BNs using the ML are much worse than those of other methods when the sample size is small and the class variable has numerous parents. To resolve the problem, we propose an exact learning augmented naive Bayes classifier (ANB), which ensures a class variable with no parents. The proposed method is guaranteed to asymptotically estimate the identical class posterior to that of the exactly learned BN. Comparison experiments demonstrated the superior performance of the proposed method.

1. Introduction

Classification contributes to solving real-world problems. The naive Bayes classifier, in which the feature variables are conditionally independent given a class variable, is a popular classifier [1]. Initially, the naive Bayes was not expected to provide highly accurate classification, because actual data were generated from more complex systems. Therefore, the general Bayesian network (GBN) with learning by marginal likelihood (ML) as a generative model was expected to outperform the naive Bayes, because the GBN is more expressive than the naive Bayes. However, Friedman et al. [2] demonstrated that the naive Bayes sometimes outperformed the GBN using a greedy search to find the smallest minimum description length (MDL) score, which was originally intended to approximate ML. They explained the inferior performance of the MDL by decomposing it into the log likelihood (LL) term, which reflects the model fitting to training data, and the penalty term, which reflects the model complexity. Moreover, they decomposed the LL term into a conditional log likelihood (CLL) of the class variable given the feature variables, which is directly related to the classification, and a joint LL of the feature variables, which is not directly related to the classification. Furthermore, they proposed conditional MDL (CMDL), a modified MDL replacing the LL with the CLL.
Consequently, Grossman and Domingos [3] claimed that the Bayesian network (BN) minimizing CMDL as a discriminative model showed better accuracy than that maximizing ML. Unfortunately, the CLL has no closed-form equation for estimating the optimal parameters. This implies that optimizing the CLL requires a gradient descent algorithm (e.g., extended logistic regression algorithm [4]). Nevertheless, the optimization algorithm involves the reiteration of each structure candidate, which renders the method computationally expensive. To solve this problem, Friedman et al. [2] proposed an augmented naive Bayes classifier (ANB) in which the class variable directly links to all feature variables, and links among feature variables are allowed. ANB ensures that all feature variables can contribute to classification. Later, various types of restricted ANBs were proposed, such as tree-augmented naive Bayes (TAN) [2] and forest-augmented naive Bayes (FAN) [5].
Because maximization of CLL entails heavy computation, various approximation methods have been proposed to maximize it. Carvalho et al. [6] proposed the approximated CLL (aCLL), which is decomposable and computationally efficient. Grossman and Domingos [3] proposed the BNC2P, which is a greedy learning method with at most two parents per variable using the hill-climbing search by maximizing CLL while estimating parameters by maximizing LL. Mihaljević et al. [7] proposed MC-DAGGES, which reduces the space for the greedy search of BN Classifiers (BNCs) using the CLL score. These reports described that the BNC maximizing the approximated CLL performed better than that maximizing the approximated ML. Nevertheless, they did not explain why CLL outperformed ML. For large data, the classification accuracies presented by maximizing ML are expected to be comparable to those presented by maximizing CLL, because ML has asymptotic consistency. Differences between the performances of the two scores in these studies might depend on their respective learning algorithms; they were approximate learning algorithms, not exact ones.
Recent studies have explored efficient algorithms for the exact learning of GBN to maximize ML [8,9,10,11,12,13,14,15,16].
This study compares the classification performances of the BNC with exact learning using ML as a generative model and those with approximate learning using CLL as a discriminative model. The results show that maximizing ML shows better classification accuracy when compared with maximizing CLL for large data. However, the results also show that classification accuracies obtained by exact learning BNC using ML are much worse than those obtained by other methods when the sample size is small and the class variable has numerous parents in the exactly learned networks. When a class variable has numerous parents, estimation of the conditional probability parameters of the class variable become unstable because the number of parent configurations becomes large and the sample size for learning the parameters becomes sparse.
To solve this problem, this study proposes an exact learning ANB which maximizes ML and ensures that the class variable has no parents. In earlier studies, the ANB constraint was used to learn the BNC as a discriminative model. In contrast, we use the ANB constraint to learn the BNC as a generative model. The proposed method asymptotically learns the optimal ANB, which asymptotically represents the true probability distribution with the fewest parameters among all possible ANB structures. Moreover, the proposed ANB is guaranteed to asymptotically estimate the identical conditional probability of the class variable to that of the exactly learned GBN. Furthermore, learning ANBs has lower computational costs than learning GBNs. Although the main theorem assumes that all feature variables are included in the Markov blanket of the class variable, this assumption does not necessarily hold. To address this problem, we propose a feature selection method using Bayes factor for exact learning of the ANB so as to avoid increasing the computational costs. Comparison experiments show that our method outperforms the other methods.

2. Background

In this section, we introduce the notation and background material required for our discussion.

2.1. Bayesian Network

A BN is a graphical model that represents conditional independence among random variables as a directed acyclic graph (DAG). The BN provides a good approximation of the joint probability distribution because it decomposes the distribution exactly into a product of the conditional probability for each variable.
Let V = X 0 , X 1 , , X n be a set of discrete variables, where X i , ( i = 0 , , n ) can take values in the set of states 1 , , r i . One can say X i = k when X i takes the state k. According to the BN structure G, the joint probability distribution is represented as
P ( X 0 , X 1 , , X n G ) = i = 0 n P ( X i Pa i G , G ) ,
where Pa i G is the parent variable set of X i in G. When the structure G is obvious from the context, we use Pa i to denote the parents. Let θ i j k be a conditional probability parameter of X i = k when the j-th instance of the parents of X i is observed (we can say Pa i = j ). Then, we define Θ i j = k = 1 r i { θ i j k } , Θ = i = 0 n j = 1 q Pa i { Θ i j } , where q Pa i = v : X v Pa i r v . A BN is a pair B = ( G , Θ ) .
The BN structure represents conditional independence assertions in the probability distribution by d-separation. First, we define collider, for which we need to define the d-separation. Letting path denote a sequence of adjacent variables, the collider is defined as follows.
Definition 1.
Assuming we have a structure G = ( V , E ) , a variable Z V on a path ρ is a collider if and only if there exist two distinct incoming edges into Z from non-adjacent variables.
We then define d-separation as explained below.
Definition 2.
Assuming we have a structure G = ( V , E ) , X , Y V , and Z V { X , Y } , the two variables X and Y are d-separated, given Z in G, if and only if every path ρ between X and Y satisfies either of the following two conditions:
  • Z includes a non-collider on ρ.
  • There is a collider Z on ρ; Z does not include Z and its descendants.
We denote the d-separation between X and Y given Z in the structure G as D s e p G ( X , Y Z ) . Two variables are d-connected if they are not d-separated.
If we have X , Y , Z V , and X and Y are not adjacent, then the following three possible types of connections characterize the d-separations: serial connections such as X Z Y , divergence connections such as X Z Y , and convergence connections such as X Z Y . The following theorem of d-separations for these connections holds.
Theorem 1
(Koller and Friedman [17]). First, assume a structure G = ( V , E ) , X , Y , Z V . If G has a convergence connection X Z Y , then the following two propositions hold:
  • Z V { X , Y , Z } , ¬ D s e p G ( X , Y Z , Z ) ,
  • Z V { X , Y , Z } , D s e p G ( X , Y Z ) .
If G has a serial connection X Z Y or divergence connection X Z Y , then negations of the above two propositions hold.
The two DAGs are Markov equivalent when they have the same d-separations.
Definition 3.
Let G 1 = ( V , E 1 ) and G 2 = ( V , E 2 ) be the two DAGs; then G 1 and G 2 are called Markov equivalent if the following holds:
X , Y V , Z V { X , Y } , D e s p G 1 ( X , Y Z ) D s e p G 2 ( X , Y Z ) .
Verma and Pearl [18] described the following theorem to identify Markov equivalence.
Theorem 2
(Verma and Pearl [18]). Two DAGs are Markov equivalent if and only if they have identical links (edges without direction) and identical convergence connections.
Let I P * ( X , Y Z ) denote that X and Y are conditionally independent given Z in the true joint probability distribution P * . A BN structure G is an independence map (I-map) if all the d-separations in G are entailed by conditional independences in P * :
Definition 4.
Assuming the true joint probability distribution P * of the random variables in a set V and a structure G = ( V , E ) , then G is an I-map if the following proposition holds:
X , Y V , Z V { X , Y } , D s e p G ( X , Y Z ) I P * ( X , Y Z ) .
Probability distributions represented by an I-map converge to P * when the sample size becomes sufficiently large.
We introduce the following notations required for our discussion on learning BNs. Let D = { x 1 , , x d , , x N } be a complete dataset consisting of N i.i.d. instances, where each instance x d is a data-vector ( x 0 d , x 1 d , , x n d ) . For a variable set Z V , we define N j Z as the number of samples of Z = j in the entire dataset D, and we define N i j k Z as the number of samples of X i = k when Z = j in D. In addition, we define a joint frequency table J F T ( Z ) and a conditional frequency table C F T ( X i , Z ) , respectively, as a list of N j Z for j = 1 , , q Z and that of N i j k Z for i = 0 , , n , j = 1 , , q Z , and k = 1 , , r i .
The likelihood of BN B, given D, is represented as
P ( D B ) = d = 1 N P ( x 0 d , x 1 d , , x n d B ) = i = 0 n j = 1 q Pa i k = 1 r i θ i j k N i j k Pa i ,
where P ( x 0 d , x 1 d , , x n d B ) represents P ( X 0 = x 0 d , X 1 = x 1 d , , X n = x n d B ) . The maximum likelihood estimators of θ i j k are given as
θ ^ i j k = N i j k Pa i N j Pa i .
The most popular parameter estimator of BNs is the expected a posteriori (EAP) of Equation (1), which is the expectation of θ i j k with respect to the density p ( Θ i j D , G ) of Equation (2), assuming Dirichlet prior density p ( Θ i j G ) of Equation (3).
θ ^ i j k = E ( θ i j k D , G ) = θ i j k · p ( Θ i j D , G ) d Θ i j = N i j k + N i j k Pa i N i j + N j Pa i .
p ( Θ i j D , G ) = Γ ( k = 1 r i ( N i j k + N i j k Pa i ) ) k = 1 r i Γ ( N i j k + N i j k Pa i ) k = 1 r i θ i j k N i j k + N i j k Pa i 1 .
p ( Θ i j G ) = Γ ( k = 1 r i N i j k ) k = 1 r i Γ ( N i j k ) k = 1 r i θ i j k N i j k 1 .
In Equations (1)–(3), N i j k denotes the hyperparameters of the Dirichlet prior distributions ( N i j k is a pseudo-sample corresponding to N i j k Pa i ), with N i j = k = 1 r i N i j k .
The BN structure must be estimated from observed data because it is generally unknown. To learn the I-map with the fewest parameters, we maximize the score with an asymptotic consistency defined as shown below.
Definition 5
(Chickering [19]). Let G 1 = ( V , E 1 ) and G 2 = ( V , E 2 ) be the structures. A scoring criterion S c o r e has an asymptotic consistency if the following two properties hold when the sample size is sufficiently large.
  • If G 1 is an I-map and G 2 is not an I-map, then S c o r e ( G 1 ) > S c o r e ( G 2 ) .
  • If G 1 and G 2 both are I-maps, and if G 1 has fewer parameters than G 2 , then S c o r e ( G 1 ) > S c o r e ( G 2 ) .
The ML score P ( D G ) is known to have asymptotic consistency [19].
When we assume the Dirichlet prior density of Equation (3), ML is represented as
P ( D G ) = i = 0 n j = 1 q Pa i Γ ( N i j ) Γ ( N i j + N j Pa i ) k = 1 r i Γ ( N i j k + N i j k Pa i ) Γ ( N i j k ) .
In particular, Heckerman et al. [20] presented the following constraint related to the hyperparameters N i j k for ML satisfying the score-equivalence assumption, where it takes the same value for the Markov equivalent structures:
N i j k = N P ( X i = k , Pa i = j G h ) ,
where N is the equivalent sample size (ESS) determined by users, and G h is the hypothetical BN structure that reflects a user’s prior knowledge. This metric was designated as the Bayesian Dirichlet equivalent (BDe) score metric. As Buntine [21] described, N i j k = N / ( r i q Pa i ) is regarded as a special case of the BDe score. Heckerman et al. [20] called this special case the Bayesian Dirichlet equivalent uniform (BDeu), defined as
P ( D G ) = i = 0 n j = 1 q Pa i Γ ( N / q Pa i ) Γ ( N / q Pa i + N j Pa i ) k = 1 r i Γ ( N / ( r i q Pa i ) + N i j k Pa i ) Γ ( N / ( r i q Pa i ) ) .
In addition, the minimum description length (MDL) score presented in (4), which approximates the negative logarithm of ML, is often used for learning BNs.
M D L ( B D ) = log N 2 | Θ | d = 1 N log P ( x 0 d , x 1 d , , x n d B ) .
The first term of Equation (4) is the penalty term, which signifies the model complexity. The second term, LL, is the fitting term that reflects the degree of model fitting to the training data.
Both BDeu and MDL are decomposable, i.e., the scores can be expressed as a sum of local scores depending only on the conditional frequency table for one variable and its parents as follows.
S c o r e ( G ) = i = 0 n S c o r e i ( Pa i ) = i = 0 n S c o r e ( C F T ( X i , Pa i ) ) .
For example, the local score of log BDeu for C F T ( X i , Pa i ) is
S c o r e i ( Pa i ) = j = 1 q Pa i log Γ ( N / q Pa i ) Γ ( N / q Pa i + N j Pa i ) k = 1 r i log Γ ( N / ( r i q Pa i ) + N i j k Pa i ) Γ ( N / ( r i q Pa i ) ) .
The decomposable score enables an extremely efficient search for structures [10,15].

2.2. Bayesian Network Classifiers

A BNC can be interpreted as a BN for which X 0 is the class variable and X 1 , , X n are feature variables. Given an instance x = ( x 1 , , x n ) for feature variables X 1 , , X n , the BNC B infers class c by maximizing the posterior probability of X 0 as
c ^ = arg max c { 1 , , r 0 } P ( c x 1 , , x n , B ) = arg max c { 1 , , r 0 } i = 0 n j = 1 q Pa i k = 1 r i θ i j k 1 i j k = arg max c { 1 , , r 0 } j = 1 q Pa 0 k = 1 r 0 θ 0 j k 1 0 j k × i : X i C j = 1 q Pa i k = 1 r i θ i j k 1 i j k ,
where 1 i j k = 1 if X i = k and Pa i = j in the case of x , and 1 i j k = 0 otherwise. Furthermore, C is a set of children of the class variable X 0 . From Equation (6), we can infer class c given only the values of the parents of X 0 , the children of X 0 , and the parents of the children of X 0 , which comprise the Markov blanket of X 0 .
However, Friedman et al. [2] reported that BNC-minimizing MDL cannot optimize classification performance. They proposed the sole use of the following CLL of the class variable given feature variables, instead of the LL for learning BNC structures.
C L L ( B D ) = d = 1 N log P ( x 0 d x 1 d , , x n d , B ) = d = 1 N log P ( x 0 d , x 1 d , , x n d B ) d = 1 N log c = 1 r 0 P ( c , x 1 d , , x n d B ) .
Furthermore, they proposed conditional MDL (CMDL), which is a modified MDL replacing LL with CLL, as shown below.
C M D L ( B D ) = log N 2 | Θ | C L L ( B D ) .
Consequently, they claimed that the BN minimizing CMDL as a discriminative model showed better accuracy than that maximizing ML as a generative model.
Unfortunately, the CLL is not decomposable, because we cannot describe the second term of Equation (7) as a sum of the log parameters in Θ . This finding implies that no closed-form equation exists for the maximum CLL estimator for Θ . Therefore, learning the network structure that minimizes the CMDL requires a search method such as gradient descent over the space of parameters for each structure candidate. Therefore, exact learning network structures by minimizing CMDL is computationally infeasible.
As a simple means of resolving that difficulty, Friedman et al. [2] proposed an ANB that ensures an edge from the class variable to each feature variable and allows edges among feature variables. Furthermore, they proposed TAN in which the class variable has no parent, and each feature variable has a class variable and at most one other feature variable as a parent variable.
Various approximate methods to maximize CLL have been proposed. Carvalho et al. [6] proposed an aCLL score, which is decomposable and computationally efficient. Let G A N B be an ANB structure. In addition, let N i j c k be the number of samples of X i = k when X 0 = c and Pa i { X 0 } = j ( i = 1 , , n ; j = 1 , , q Pa i { X 0 } ; c = 1 , , r 0 ; k = 1 , , r i ) . In addition, let N > 0 represent the number of pseudocounts. Under several assumptions, the aCLL can be represented as   
a C L L ( G A N B D ) i = 1 n j = 1 q Pa i { X 0 } k = 1 r i c = 1 r 0 N i j c k + β c = 1 r 0 N i j c k log N i j + c k N i j + c ,
where
N i j + c k = N i j c k + β c = 1 r 0 N i j c k if N i j c k + β c = 1 r 0 N i j c k N N otherwise ,
N i j + c = k = 1 r i N i j + c k .
The value of β is found by using the Monte Carlo method to approximate the CLL. When the value of β is optimal, the aCLL is a minimum-variance unbiased approximation of the CLL.
Moreover, Grossman and Domingos [3] proposed a learning structure method using a greedy hill-climbing algorithm [20] by maximizing the CLL while estimating the parameters by maximizing the LL. Recently, Mihaljević et al. [7] identified the smallest subspace of DAGs that covered all possible class-posterior distributions when the data were complete.
All the DAGs in this space, which they call minimal class-focused DAGs (MC-DAGs), are such that every edge is directed toward a child of the class variable. In addition, they proposed a greedy search algorithm in the space of Markov equivalent classes of MC-DAGs using the CLL score. These reports described that the BNC maximizing the approximated CLL provides better performance than that maximizing the approximated ML. However, they did not explain why CLL outperformed ML. For large data, the classification accuracies obtained by maximizing ML are expected to be comparable to those obtained by maximizing CLL because ML has asymptotic consistency. Differences between the performances of the two scores in these earlier studies might depend on their learning algorithms to maximize ML; they were approximate learning algorithms, not exact ones.

3. Classification Accuracies of Exact Learning GBN

This section presents experiments comparing the classification accuracies of the exactly learned GBN by maximizing the BDeu as a generative model with those of the approximately learned BNC by maximizing the CLL as a discriminative model. Although determining the hyperparameter N of BDeu is difficult [16,22,23,24], we use N = 1.0 , which allows the data to reflect the estimated parameters to the greatest degree possible [25,26].
The experiment compared the respective classification accuracies of seven methods in Table 1. All the methods were implemented in Java. The source code is available at http://www.ai.lab.uec.ac.jp/software/ (accessed on 29 November 2021). Throughout this paper, our experiments were conducted on a computational environment, as shown in Table 2. This experiment used 43 classification benchmark datasets from the UCI repository [27]. Continuous variables were discretized into two bins using the median value as the cutoff, as in [28]. In addition, data with missing values were removed from the datasets. We used EAP estimators as conditional probability parameters of the respective classifiers. The hyperparameters N i j k of EAP were found to be 1 / ( r i q Pa i ) . Throughout our experiments, we defined “small datasets” as the datasets with less than 200 samples, and we defined “large datasets” as the datasets with 10,000 or more samples.
Table 3 presents the classification accuracies of the respective classifiers. However, we will discuss the results of ANB-BDeu and fsANB-BDeu in a later section. The values shown in bold in Table 3 represent the best classification accuracies for each dataset. Here, the classification accuracies represent the average percentage of correct classifications from a ten-fold cross-validation. Moreover, to investigate the relation between the classification accuracies and GBN-BDeu, Table 4 presents the details of the achieved structures using GBN-BDeu. “Parents” in Table 4 represents the average number of the class variable’s parents in the structures learned by GBN-BDeu. “Children” denotes the average number of the class variable’s children in the structures learned by GBN-BDeu. “Sparse data” denotes the average number of patterns of X 0 ’s parents value j with null data, N j Pa 0 = 0 ( j = 1 , , q Pa 0 ) in the structures learned by GBN-BDeu.
From Table 3, GBN-BDeu shows the best classification accuracies among the methods for large data, such as dataset Nos. 22, 29, and 33. Because BDeu has asymptotic consistency, the joint probability distribution represented by GBN-BDeu approaches the true distribution as the sample size increases. However, it is worth noting that GBN-BDeu provides much worse accuracy than the other methods in datasets No. 3 and No. 9. In these datasets, the learned class variables by GBN-BDeu have no children. Numerous parents are shown in “Parents” and “Children” in Table 4. When a class variable has numerous parents, the estimation of the conditional probability parameters of the class variable becomes unstable, because the class variable’s parent configurations become numerous. Then, the sample size for learning the parameters becomes small, as presented in “Sparse data” in Table 4. Therefore, numerous parents of the class variable might be unable to reflect the feature data for classification when the sample is insufficiently large.

4. Exact Learning ANB Classifier

The preceding section suggested that exact learning of GBN by maximizing BDeu to have no parents of the class variable might improve the accuracy of GBN-BDeu. In this section, we propose an exact learning ANB, which maximizes BDeu and ensures that the class variable has no parents. In earlier reports, the ANB constraint was used to learn the BNC as a discriminative model. In contrast, we use the ANB constraint to learn the BNC as a generative model. The space of all possible ANB structures includes at least one I-map, because it includes a complete graph, which is an I-map. From the asymptotic consistency of BDeu (Definition 5), the proposed method is guaranteed to achieve the I-map with the fewest parameters among all possible ANB structures when the sample size becomes sufficiently large. Our empirical analysis in Section 3 suggests that the proposed method can improve the classification accuracy for small data. We employed the dynamic programming (DP) algorithm learning GBN [10] for the exact learning of ANB. The DP algorithm for exact learning ANB was almost twice as fast as that for the exact learning of GBN. We prove that the proposed ANB asymptotically estimates the identical conditional probability of the class variable to that of the exactly learned GBN.

4.1. Learning Procedure

The proposed method is intended to seek the optimal structure that maximizes the BDeu score among all possible ANB structures. The local score of the class variable in ANB structures is constant because the class variable has no parents in the ANB structure. Therefore, we can ascertain the optimal ANB structure by maximizing S c o r e A N B ( G ) = S c o r e ( G ) S c o r e 0 ( ϕ ) .
Before we describe the procedure of our method, we introduce the following notations. Let G * ( Z ) denote the optimal ANB structure composed of a variable set Z , ( X 0 Z ) . When a variable has no child in a structure, we say it is a sink in the structure. We use X s * ( Z ) to denote a sink in G * ( Z ) . Additionally, letting Π ( Z ) denote a set of all the Z ’s subsets including X 0 , we define the best parents of X i in a candidate set Π ( Z ) , as the parent set that maximizes the local score in Π ( Z ) :
g i * ( Π ( Z ) ) = arg max W Π ( Z )   S c o r e i ( W ) .
Our algorithm has four logical steps. The following process improves the DP algorithm proposed by [10] to learn the optimal ANB structure.
(1)
For all possible pairs of a variable X i V { X 0 } and a variable set Z V { X i } , ( X 0 Z ) , calculate the local score S c o r e i ( Z ) (Equation (5)).
(2)
For all possible pairs of a variable X i V { X 0 } and a variable set Z V { X i } , ( X 0 Z ) , calculate the best parents g * ( Π ( Z ) ) .
(3)
For Z V , ( X 0 Z ) , calculate the sink X s * ( Z ) .
(4)
Calculate G * ( V ) using Steps 3 and 4.
Steps 3 and 4 of the algorithm are based on the observation that the best network G * ( Z ) necessarily has a sink X s * ( Z ) with incoming edges from its best parents g s * ( Π ( Z { X s * ( Z ) } ) ) . The remaining variables and edges in G * ( Z ) necessarily construct the best network G * ( Z { X s * ( Z ) } ) . More formally,
X s * ( Z ) = arg max X i Z { X 0 } S c o r e i ( g i * ( Π ( Z { X i } ) ) ) + S c o r e A N B ( G * ( Z { X i } ) ) .
From Equation (8), we can decompose G * ( Z ) into G * ( Z { X s * ( Z ) } ) and X s * ( Z ) with incoming edges from g s * ( Π ( Z { X s * ( Z ) } ) . Moreover, this decomposition can be performed recursively. At the end of the recursive decomposition, we obtain n pairs of the sink and its best network, denoted by ( X s 1 , g s 1 * ) , , ( X s i , g s i * ) , , ( X s n , g s n * ) . Finally, we obtain G * ( V ) for which X s i ’s parent set is g s i * .
The number of iterations to calculate all the local scores, best parents, and best sinks for our algorithm are ( n 1 ) 2 n 2 , ( n 1 ) 2 n 2 , and 2 n 1 , respectively, and those for GBN are n 2 n 1 , n 2 n 1 , and 2 n , respectively. Therefore, the DP algorithm for ANB is almost twice as fast as that for GBN. The details of the proposed algorithm are shown in the Appendix A.

4.2. Asymptotic Properties of the Proposed Method

Under some assumptions, the proposed ANB is proven to asymptotically estimate the identical conditional probability of the class variable, given the feature variables of the exactly learned GBN. When the sample size becomes sufficiently large, the structure learned by the proposed method and the exactly learned GBN are classification-equivalent defined as follows:
Definition 6
(Acid et al. [29]). Let G be a set of all the BN structures. Furthermore, let D be any finite dataset. For G 1 , G 2 G , we say that G 1 and G 2 are classification-equivalent if P ( X 0 x , G 1 , D ) = P ( X 0 x , G 2 , D ) for any feature variable’s value x .
To derive the main theorem, we introduce five lemmas as below.
Lemma 1
(Mihaljević et al. [7]). Let G = ( V , E ) be a structure. Then, G is classification-equivalent to G , which is a modified G by the following operations:
(1)
For X , Y Pa 0 G , add an edge between X and Y in G.
(2)
For X Pa 0 G , reverse an edge from X to X 0 in G.
Next, we use the following lemma from Chickering [19] to derive the main theorem:
Lemma 2
(Chickering [19]). Let G I m a p be a set of all I-maps. When the sample size becomes sufficiently large, then the following proposition holds.
G 1 , G 2 G I m a p , ( ( X , Y V , Z V { X , Y } , D s e p G 1 ( X , Y Z ) D s e p G 2 ( X , Y Z ) ) S c o r e ( G 1 ) S c o r e ( G 2 ) ) .
Moreover, we provide Lemma 3 under the following assumption.
Assumption 1.
Let the true joint probability distribution of random variables in a set V be P * . Under Assumption A1, a true structure G * = ( V , E * ) exists that satisfies the following property:
X , Y V , Z V { X , Y } , D s e p G * ( X , Y Z ) I P * ( X , Y Z ) .
Lemma 3.
Let G A N B I m a p be a set of all the ANB structures that are I-maps. For G A N B I m a p G A N B I m a p , X , Y V , if G * has a convergence connection X X 0 Y , then G A N B I m a p has an edge between X and Y.
Proof. 
We prove Lemma 3 by contradiction. Assuming that G A N B I m a p has no edge between X and Y, because G A N B I m a p has a divergence connection X X 0 Y , we obtain
Z V { X , Y , X 0 } , D s e p G A N B I m a p ( X , Y X 0 , Z ) .
Because G * has a convergence connection X X 0 Y , the following proposition holds from Theorem 1:
Z V { X , Y , X 0 } , ¬ D s e p G A N B I m a p ( X , Y X 0 , Z ) .
This result contradicts (9). Consequently, G A N B I m a p has an edge between X and Y.    □
Furthermore, under Assumption A1 and the following assumptions, we derive Lemma 4.
Assumption 2.
All feature variables are included in the Markov blanket M of the class variable in the true structure G * .
Assumption 3.
For X M , X and X 0 are adjacent to G * .
Lemma 4.
Let G 1 * be the modified G * by operation 1 in Lemma 1. In addition, let G 12 * be the structure that is modified to G 1 * by operation 2 in Lemma 1. Under Assumptions 1–3, G 1 * is Markov equivalent to G 12 * .
Proof. 
From Theorem 2, we prove Lemma 4 by showing the following two propositions: (I)  G 1 * and G 12 * have the same links (edges without direction), and (II) they have the same set of convergence connections. Proposition (I) can be proved immediately because the difference between G 1 * and G 12 * is only the direction of the edges between X 0 and the variables in Pa 0 G * . For the same reason, G 1 * and G 12 * have the same set of convergence connections as colliders in V ( Pa 0 G * { X 0 } ) . Moreover, there are no convergence connections with colliders in Pa 0 G * { X 0 } in both G 1 * and G 12 * because all the variables in Pa 0 G * { X 0 } are adjacent in the two structures. Consequently, they have the same set of convergence connections, so Proposition (II) holds. This completes the proof.    □
Finally, under Assumptions 1–3, we derive the following lemma.
Lemma 5.
Under Assumptions 1–3, G 12 * is an I-map.
Proof. 
The DAG G 1 * results from adding the edges between the variables in Pa 0 G * to G * . Because adding edges does not create a new d-separation, G 1 * remains an I-map. Lemma 5 holds because G 1 * is a Markov equivalent to G 12 * from Lemma 4.    □
Under Assumptions 1–3, we prove the following main theorem using Lemmas 1–5.
Theorem 3.
Under Assumptions 1–3, when the sample becomes sufficiently large, the proposal (learning ANB using BDeu) achieves the classification-equivalent structure to G * .
Proof. 
Because G 12 * is classification-equivalent to G * from Lemma 1, we prove Theorem 3 by showing that the proposed method asymptotically learns a Markov-equivalent structure to G 12 * . We prove Theorem 3 by showing that G 12 * asymptotically has the maximum BDeu score among all the ANB structures:
G A N B G A N B , S c o r e ( G A N B ) S c o r e ( G 12 * ) .
From Definition 5, the BDeu scores of the I-maps are higher than those of any non-I-maps when the sample size becomes sufficiently large. Therefore, it is sufficient to show that the following proposition holds asymptotically to prove that Proposition (11) asymptotically holds.
G A N B I m a p G A N B I m a p , S c o r e ( G A N B I m a p ) S c o r e ( G 12 * ) .
From Lemma 5, G 12 * is an I-map. Therefore, from Lemma 2, a sufficient condition for satisfying (12) is as follows:
G A N B I m a p G A N B I m a p , X , Y M { X 0 } , Z M { X 0 } { X , Y } , D s e p G A N B I m a p ( X , Y Z ) D s e p G 12 * ( X , Y Z ) .
We prove (13) by dividing it into two cases: X Pa 0 G * Y Pa 0 G * and X Pa 0 G * Y Pa 0 G * .
  • Case I: X Pa 0 G * Y Pa 0 G *
  •    From Lemma 3, all variables in Pa 0 G * are adjacent to G A N B I m a p . Therefore, we obtain
    Z M { X 0 } { X , Y } , ¬ D s e p G A N B I m a p ( X , Y Z ) ¬ D s e p G 12 * ( X , Y Z ) .
  •    For two Boolean propositions p and q, the following holds.
    ( ¬ p ¬ q ) ( p q )
  •    From (14) and (15), we obtain
    Z M { X 0 } { X , Y } , D s e p G A N B I m a p ( X , Y Z ) D s e p G 12 * ( X , Y Z ) .
  • Case II: X Pa 0 G * Y Pa 0 G *
  •    From Definition 4 and Assumption A1, we obtain
    Z M { X 0 } { X , Y } , D s e p G A N B I m a p ( X , Y Z ) D s e p G * ( X , Y Z ) .
  •    Thus, we can prove (13) by showing that the following proposition holds:
    Z M { X 0 } { X , Y } , D s e p G * ( X , Y Z ) D s e p G 12 * ( X , Y Z ) .
  • For the remainder of the proof, we prove the sufficient condition (16) to satisfy (13) by dividing it into two cases: X 0 Z and X 0 Z .
  •    Case i: X 0 Z
    • All pairs of non-adjacent variables in Pa 0 G * in G * comprise a convergence connection with collider X 0 . From Theorem 1, these pairs are necessarily d-connected, given X 0 in G * . Therefore, all the variables in Pa 0 G * are d-connected, given X 0 in G * . This means that G * and G 1 * represent identical d-separations given X 0 . Because G 1 * is Markov equivalent to G 12 * from Lemma 4, G * and G 12 * represent identical d-separations given X 0 ; i.e., Proposition (16) holds.
  •    Case ii: X 0 Z
  •       We divide (16) into two cases: X = X 0 Y = X 0 and X X 0 Y X 0
  •       Case 1: X = X 0 Y = X 0
    • Because all the variables in the X 0 ’s Markov blanket M are adjacent to X 0 in both G 12 * and G * from Assumption A2, we obtain ¬ D s e p G 12 * ( X , Y Z ) ¬ D s e p G * ( X , Y Z ) . From (15), proposition (16) holds.
  •       Case 2: X X 0 Y X 0
    • If both G 12 * and G * have no edge between X and Y, they have a serial or divergence connection: X X 0 Y or X X 0 Y . Because the serial and divergence connections represent d-connections between X and Y in this case from Theorem 1, we obtain ¬ D s e p G 12 * ( X , Y Z ) ¬ D s e p G * ( X , Y Z ) . From (15), proposition (16) holds.
  •    Thus, we complete the proof of (13) in Case II.
Consequently, proposition (13) is true, which completes the proof of Theorem 3.    □
We proved that the proposed ANB asymptotically estimates the identical conditional probability of the class variable to that of the exactly learned GBN.

4.3. Numerical Examples

This subsection presents the numerical experiments conducted to demonstrate the asymptotic properties of the proposed method. To demonstrate that the proposed method asymptotically achieves the I-map with the fewest parameters among all the possible ANB structures, we evaluated the structural Hamming distance (SHD) [30], which measures the distance between the structure learned by the proposed method and the I-map with the fewest parameters among all the possible ANB structures. To demonstrate Theorem 3, we evaluated the Kullback–Leibler divergence (KLD) between the learned class variable posterior using the proposed method and that by the true structure. This experiment used two benchmark datasets from bnlearn [31]: CANCER and ASIA, as depicted in Figure 1 and Figure 2. We used the variables “Cancer” and “either” as the class variables in CANCER and ASIA, respectively. In that case, CANCER satisfied Assumptions 2 and 3, but ASIA did not.
From the two networks, we randomly generated sample data for each sample size N = 100, 500, 1000, 5000, 10,000, 50,000, and 100,000. Based on the generated data, we learned the BNC structures using the proposed method and then evaluated the SHDs and KLDs. Table 5 presents the results. The results show that the SHD converged to 0 when the sample size increased in both CANCER and ASIA. Thus, the proposed method asymptotically learned the I-map with the fewest parameters among all possible ANB structures. Furthermore, in CANCER, the KLD between the learned class variable posterior by the proposed method and that by the true structure became 0 when N 1000. The results demonstrate that the proposed method learns the classification-equivalent structure of the true one when the sample size becomes sufficiently large, as described in Theorem 3. In ASIA however, the KLD between the learned class variable posterior by the proposed method and that by the true structure did not reach 0 even when the sample size became large because ASIA did not satisfy Assumptions 2 and 3.

5. Learning Markov Blanket

Theorem 3 assumes all feature variables are included in the Markov blanket of the class variable. However, this assumption does not necessarily hold. To solve this problem, we must learn the Markov blanket of the class variable before learning the ANB. Under Assumption 3, the Markov blanket of the class variable is equivalent to the parent-child (PC) set of the class variable. It is known that the exact learning of a PC set of variables is computationally infeasible when the number of variables increases. To reduce the computational cost of learning a PC set, ref. [32] proposed a score-based local learning algorithm (SLL), which has two learning steps. In step 1, the algorithm sequentially learns the PC set by repeatedly using the exact learning structure algorithm on a set of variables containing the class variable, the current PC set, and one new query variable. In step 2, SLL enforces the symmetry constraint: if X i is a child of X j , then X j is a parent of X i . This allows us to try removing extra variables from PC set, proving that the SLL algorithm always finds the correct PC of the class variable when the sample size is sufficiently large. Moreover, ref. [33] proposed the S 2 TMB algorithm, which improved the efficiency over the SLL by removing the symmetric constraints in PC search steps. However, S 2 TMB is computationally infeasible when the size of the PC set surpasses 30.
As an alternative approach for learning large PC sets, previous studies proposed constraint-based PC search algorithms, such as MMPC [30], HITON-PC [34], and PCMB [35]. These methods produce an undirected graph structure using statistical hypothesis tests or information theory tests. As statistical hypothesis tests, the G 2 and χ 2 tests were used for these constraint-based methods. In these tests, the independence of the two variables was set as a null hypothesis. A p-value signifies the probability that the null hypothesis is correct at a user-determined significance level. If the p-value exceeds the significance level, the null hypothesis is accepted, and the edge is removed. However, [36] reported that statistical hypothesis tests have a significant shortcoming: the p-value sometimes becomes much smaller than the significance level as the sample size increases. Therefore, statistical hypothesis tests suffer from Type I errors (detecting dependence for an independent conditional relation in the true DAG). Conditional mutual information (CMI) is often used as a CI test [37]. The CMI strongly depends on a hand-tuned threshold value. Therefore, it is not guaranteed to estimate the true CI structure. Consequently, CI tests have no asymptotic consistency.
For a CI test with asymptotic consistency, [38] proposed a Bayes factor with BDeu (the “BF method”, below), where the Bayes factor is the ratio of marginal likelihoods between two hypotheses [39]. For two variables X , Y V and a set of conditional variables Z V { X , Y } , the BF method B F ( X , Y Z ) is defined as
B F ( X , Y Z ) = exp ( S c o r e ( C F T ( X , Z ) ) ) exp ( S c o r e ( C F T ( X , Z { Y } ) ) ) ,
where S c o r e ( C F T ( X , Z ) ) and S c o r e ( C F T ( X , Z { Y } ) ) can be obtained using Equation (5). The BF method detects I P * ( X , Y Z ) if B F ( X , Y Z ) is larger than the threshold δ , and detects the ¬ I P * ( X , Y Z ) otherwise. Natori et al. [40] and Natori et al. [41] applied the BF method to a constraint-based approach, and showed that their method was more accurate than the other methods with traditional CI tests.
We propose the constraint-based PC search algorithm using a BF method. The proposed PC search algorithm finds the PC set of the class variable using a BF method between the class variable and all feature variables because the Bayes factor has an asymptotic consistency for the CI tests [41]. It is known that missing crucial variables degrades the accuracy [2]. Therefore, we redundantly learn the PC set of the class variable to reduce extra variables with no missing variables as follows.
  • The proposed PC search algorithm only conducts the CI tests at the zero order (given no conditional variables), which is more reliable than those at the higher order.
  • We use a positive value as Bayes factor’s threshold δ .
Furthermore, we compare the accuracy of the proposed PC search method with those of the MMPC, HITON-PC, PCMB, and S 2 TMB. Learning Bayesian networks is known to be highly sensitive to the chosen ESS-value [22,25,26]. Therefore, we determine the ESS N { 1.0 , 2.0 , 5.0 } and the threshold δ { 3 , 20 , 150 } in the Bayes factor using two-fold cross validation to obtain the highest classification accuracy. The three ESS-values of N are determined according to Ueno [25], Ueno [26]. The three values of δ are determined according to Heckerman et al. [20]. All the compared methods were implemented in Java (Source code is available at http://www.ai.lab.uec.ac.jp/software/, accessed on 29 November 2021). This experiment used six benchmark datasets from bnlearn: ASIA, SACHS, CHILD, WATER, ALARM, and BARLEY. From each benchmark network, we randomly generated sample data N = 10,000. Based on the generated data, we learned all the variables’ PC sets using each method. Table 6 shows the average runtimes of each method. We calculated missing variables, representing the number of removed variables existing in the true PC set, and extra variables, which indicated the number of remaining variables that do not exist in the true PC set. Table 6 also shows the average missing and extra variables from the learned PC sets of all the variables. We compared the classification accuracies of the exact learning ANB with BDeu score (designated as ANB-BDeu), using each PC search method as a feature selection method. Table 7 shows the average accuracies of each method from the 43 UCI repository datasets listed in Table 3.
From Table 6, the results show that the runtimes of the proposed method were shorter than those of the other methods. Moreover, the results show that the missing variables of the proposed method were smaller than those of the other methods. On the other hand, Table 6 also shows that the extra variables of the proposal were greater than those of the other methods in all datasets. From Table 7, the results show that the ANB-BDeu using the proposed method provided a much higher average accuracy than the other methods. This is because missing variables degrade classification accuracy more significantly than extra variables (Friedman et al., 1997).

6. Experiments

This section presents numerical experiments conducted to evaluate the effectiveness of the exact learning ANB. First, we compared the classification accuracies of ANB-BDeu with those of the other methods in Section 3. We used the same experimental setup and evaluation method described in Section 3. The classification accuracies of ANB-BDeu are presented in Table 3. To confirm the significant differences of ANB-BDeu from the other methods, we applied Hommel’s tests [42], which are used as a standard in machine learning studies [43]. The p-values are presented at the bottom of Table 3. In addition, “MB size” in Table 4 denotes the average number of the class variable’s Markov blanket size in the structures learned by GBN-BDeu.
The results show that ANB-BDeu outperformed Naive Bayes, GBN-CMDL, BNC2P, TAN-aCLL, gGBN-BDeu, and MC-DAGGES at the p < 0.1 significance level. Moreover, the results show that ANB-BDeu improved the accuracy of GBN-BDeu when the class variable had numerous parents such as the No. 3, No. 9, and No. 31 datasets, as shown in Table 4. Furthermore, ANB-BDeu provided higher accuracies than GBN-BDeu, even for large data such as datasets 13, 22, 29, and 33, although the difference between ANB-BDeu and GBN-BDeu was not statistically significant. These actual datasets did not necessarily satisfy Assumptions 1–3 in Theorem 3. These results imply that the accuracies of ANB-BDeu without satisfying Assumptions 1–3 might be comparable to those of GBN-BDeu for large data. It is worth noting that the accuracies of ANB-BDeu were much worse than those provided by GBN-BDeu for datasets No. 5 and No. 12. “MB size” in these datasets were much smaller than the number of all feature variables, as shown in Table 4. The results show that feature selection by the Markov blanket is expected to improve the classification accuracies of the exact learning ANB, as described in Section 5.
We compared the classification accuracies of ANB-BDeu using the PC search method proposed in Section 5 (referred to as “fsANB-BDe”) with the other methods in Table 3. Table 3 shows the classification accuracies of fsANB-BDe and the p-values of Hommel’s tests for differences in fsANB-BDeu from the other methods. The results show that fsANB-BDeu outperformed all the compared methods at the p < 0.05 significance level.
“Max parents” in Table 4 presents the average maximum number of parents learned by fsANB-BDeu. The value of “Max parents” represents the complexity of the structure learned by fsANB-BDeu. The results show that the accuracies of Naive Bayes were better than those of fsANB-BDeu when the sample size was small, such as the No. 36 and No. 38 datasets. In these datasets, the values of “Max parents” are large. The estimation of the variable parameters tends to become unstable when a variable has numerous parents, as described in Section 3. Naive Bayes can avoid this phenomenon because the maximum number of parents in Naive Bayes is one. However, Naive Bayes cannot learn relationships between the feature variables. Therefore, for large samples such as the No. 8 and No. 29 datasets, Naive Bayes showed much worse accuracy than those provided by other methods.
Similar to Naive Bayes, BNC2P and TAN-aCLL show better accuracies than fsANB-BDeu for small samples such as the No. 38 dataset because the upper bound of the maximum number of parents was two in the two methods. However, the small upper bound of the maximum number of parents tends to lead to a poor representational power of the structure [44]. As a result, the accuracies of both methods tend to be worse than those of the fsANB-BDeu of which the value of “Max parents” is greater than two, such as the No. 29 dataset.
For large samples, such as datasets Nos 29 and 33, GBN-CMDL, gGBN-BDeu, and MC-DAGGES showed worse accuracies than fsANB-BDeu, because the exact learning methods estimate the network structure more precisely than the greedy learned structure.
We compared fsANB-BDeu and ANB-BDeu. The difference between the two methods is whether the proposed PC search method is used. “Removed variables” in Table 4 represents the average number of variables removed from the class variable’s Markov blanket by our proposed PC search method. The results demonstrated that the accuracies of fsANB-BDeu tended to be much higher than those of ANB-BDeu when the value of “Removed variables” was large, such as Nos. 5, 12, 16, 34, and 38. Consequently, discarding numerous irrelevant variables in the features improved the classification accuracy.
Finally, we compared the runtimes of fsANB-BDeu and GBN-BDeu to demonstrate the efficiency of the ANB constraint. Table 8 presents the runtimes of GBN-BDeu, fsANB-BDeu, and the proposed PC search method. The results show that the runtimes of fsANB-BDeu were shorter than those of GBN-BDeu in all the datasets, because the execution speed of the exact learning ANB was almost twice that of the exact learning GBN, as described in Section 4. Moreover, the runtimes of fsANB-BDeu were much shorter than those of GBN-BDeu when our PC search method removed many variables, such as the No. 34 and No. 39 datasets. This is because the runtimes of GBN-BDeu decrease exponentially with the removal of variables, whereas our PC search method itself has a negligibly small runtime compared to those of the exact learning as shown in Table 8.
As a result, the proposed method fsANB-BDeu provides the best classification performances in all the methods with a lower computational cost than that of the GBN-BDeu.

7. Conclusions

First, this study compared the classification performances of the BNs exactly learned by BDeu as a generative model and those learned approximately by CLL as a discriminative model. Surprisingly, the results demonstrated that the performance of BNs achieved by maximizing ML was better than that of BNs achieved by maximizing CLL for large data. However, the results also showed that the classification accuracies of the BNs that learned exactly by BDeu were much worse than those that learned by the other methods when the class variable had numerous parents. To solve this problem, this study proposed an exact learning ANB by maximizing BDeu as a generative model. The proposed method asymptotically learned the optimal ANB, which is an I-map with the fewest parameters among all possible ANB structures. In addition, the proposed ANB is guaranteed to asymptotically estimate the identical conditional probability of the class variable to that of the exactly learned GBN. Based on these properties, the proposed method is effective for not only classification but also decision making, which requires a highly accurate probability estimate of the class variable. Furthermore, the learning ANB has lower computational costs than the learning BN does. The experimental results demonstrated that the proposed method significantly outperformed the approximately learned structure by maximizing CLL.
We plan on exploring the following in future work.
(1)
It is known that neural networks are universal approximators, which means that they can approximate any functions to an arbitrary small error. However, Choi et al. [45] showed that the functions induced by BN queries are polynomials. To improve their queries to become universal approximators, they proposed a testing BN, which chooses a parameter value depending on a threshold instead of simply having a fixed parameter value. We will apply our proposed method to the testing BN.
(2)
Recent studies have developed methods for compiling BNCs into Boolean circuits that have the same input–output behavior [46,47]. We can explain and verify any BNCs by operating on their compiled circuits [47,48,49]. We will apply the compiling method to our proposed method.
(3)
Sugahara et al. [50] proposed the Bayesian network model averaging classifier with subbagging to improve the classification accuracy for small data. We will extend our proposed method to the model averaging classifier.
The above future works are expected to improve the classification accuracies and comprehensibility of our proposed method.

Author Contributions

Conceptualization, methodology, S.S. and M.U.; validation, S.S. and M.U.; writing—original draft preparation, S.S.; writing—review and editing, M.U.; funding acquisition, M.U. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by JSPS KAKENHI Grant Numbers JP19H05663 and JP19K21751.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The datasets used in our experiments are available at: http://www.ai.lab.uec.ac.jp/software/ (accessed on 29 November 2021).

Acknowledgments

Parts of this research were reported in an earlier conference paper published by Sugahara et al. [51].

Conflicts of Interest

The authors declare that they have no conflict of interest. The funders had no role in the design of the study, in the collection, analyses, or interpretation of data, in the writing of the manuscript, or in the decision to publish the results.

Appendix A

In this section, we provide the detailed algorithm of the exact learning ANB with BDeu score. As described in Section 4, our algorithm has the following steps.
(1)
For all possible pairs of a variable X i V { X 0 } and a variable set Z V { X i } , ( X 0 Z ) , calculate the local score S c o r e i ( Z ) (Equation (5)).
(2)
For all possible pairs of a variable X i V { X 0 } and a variable set Z V { X i } , ( X 0 Z ) , calculate the best parents g * ( Π ( Z ) ) .
(3)
For Z V , ( X 0 Z ) , calculate the sink X s * ( Z ) .
(4)
Calculate G * ( V ) using Steps 3 and 4.
Although our algorithm employs the approach provided by [10], the main differences are that our algorithm does not calculate the local scores of the parent sets without X 0 in Step 1 and does not search these parent sets in Step 2. Hereinafter, we explain how steps 1–4 can be accomplished within a reasonable time.
First, we calculate the joint frequency table of the entire variable set V , and calculate the joint frequency tables of smaller variable subsets by marginalizing a variable, except X 0 , from the joint frequency table. Using each joint frequency table, we calculate the conditional frequency tables for each variable, except X 0 , given the other variables in the joint frequency table. We use these conditional frequency tables to calculate the local log BDeu scores. This process calculates the local scores only for the parent sets, including X 0 , which satisfies the ANB constraint.
We call G e t L o c a l S c o r e s described in Algorithm A1 with a joint frequency table j f t and the variables e f v s , which are marginalized from j f t recursively. By initially calling G e t L o c a l S c o r e s with a joint frequency table for V and the variable set V { X 0 } as e f v s , we can calculate all the local scores required in Step 1. The algorithm calculates increasingly smaller joint frequency tables by a depth-first search.
Algorithm A1 G e t L o c a l S c o r e s ( j f t , e f v s )
  • for all X F v s ( j f t ) do
  •      L S [ X ] [ F v s ( j f t ) { X 0 } { X } ] S c o r e ( J f t 2 c f t ( j f t , X ) )
  • end for
  • if | F v s ( c t ) | > 1 then
  •     for  j = 1 to | e f v s |  do
  •          G e t L o c a l S c o r e s ( J f t 2 j f t ( j f t , X i ) , { e f v s [ 1 ] , , e f v s [ j 1 ] } )
  •     end for
  • end if
Algorithm A1 uses the following subfunction: F v s ( j f t ) returns the set of variables except X 0 in the joint frequency table j f t ; J f t 2 j f t ( j f t , X ) yields a joint frequency table, with the variable X marginalized, from j f t ; J f t 2 c f t ( j f t , X ) produces a conditional frequency table. The calculated scores for each (variable, parent set) pair are stored in L S .
After Step 1, we can find the best parents recursively from the calculated local scores. For a variable set Z V , ( X 0 Z ) , the best parents of X i V { X 0 } in a candidate set Π ( Z ) are either Π ( Z ) itself or the best parents of X i in one of the smaller candidate sets { Π ( Z { X } ) X ( Z { X 0 } ) } . More formally, one can say that
S c o r e i ( g i * ( Π ( Z ) ) ) = max ( S c o r e i ( Z ) , S c o r e 1 ( Z ) ) ,
where
S c o r e 1 ( Z ) = max X Z { X 0 } S c o r e i ( g i * ( Π ( Z { X } ) ) ) .
Using this relation, Algorithm A2 finds all the best parents required in Step 2 by calculating the Formula (A1) in the lexicographic order of the candidate sets. The algorithm is called with the variable X i V { X 0 } , the variable set V , and the previously calculated local scores L S . The identified best parents and their local scores are stored in b p s and b s s , respectively.
Algorithm A2 G e t B e s t P a r e n t s ( V , X i , L S )
  • b p s = array 1 to 2 n 2 of variable sets
  • b s s = array 1 to 2 n 2 of local scores
  • for all c s ( V { X i } ) such that X 0 c s in lexicographic order do
  •      b p s [ c s ] c s
  •      b s s [ c s ] L S [ X i ] [ c s ]
  •     for all  c s 1 c s such that X 0 c s 1 and | c s c s 1 | = 1  do
  •         if  b s s [ c s 1 ] > b s s [ c s ]  then
  •             b s s [ c s ] b s s [ c s 1 ]
  •             b p s [ c s ] b p s [ c s 1 ]
  •         end if
  •     end for
  • end for
As described previously, the best network G * ( Z ) can be decomposed into the smaller best network G * ( Z { X s * ( Z ) } ) and the best sink X s * ( Z ) with incoming edges from g s * ( Π ( Z { X s * ( Z ) } ) . Again, using this idea, Algorithm A3 finds all the best sinks required in Step 3 by calculating the Formula (8) in the lexicographic order of the variable sets. The identified best sinks are stored in s i n k s .
Algorithm A3 G e t B e s t S i n k s ( V , b p s , L S )
  • for all Z V such that X 0 Z in lexicographic order do
  •      s c o r e s [ Z ] 0.0
  •      s i n k s [ Z ] 1
  •     for all  s i n k Z { X 0 }  do
  •          u p v a r s Z { s i n k }
  •          s k o r e s c o r e s [ u p v a r s ]
  •          s k o r e s k o r e + L S [ s i n k ] [ b p s [ s i n k ] [ u p v a r s ] ]
  •         if  s i n k s [ Z ] = 1 or s k o r e > s c o r e s [ Z ]  then
  •             s c o r e s [ Z ] s k o r e
  •             s i n k s [ Z ] s i n k
  •         end if
  •     end for
  • end for
At the end of the recursive decomposition using Equation (8), we can identify the best network G * ( V ) , as described in Algorithm A4.
Algorithm A4 G e t B e s t N e t ( V , b p s , s i n k s )
  • p a r e n t s G * = array 1 to n of variable sets
  • l e f t = V
  • for i = 1 tondo
  •      X s s i n k [ l e f t ]
  •      l e f t l e f t { X s }
  •      p a r e n t s G * [ X s ] b p s [ X s ] [ l e f t ]
  • end for

References

  1. Minsky, M. Steps toward Artificial Intelligence. Proc. IRE 1961, 49, 8–30. [Google Scholar] [CrossRef]
  2. Friedman, N.; Geiger, D.; Goldszmidt, M. Bayesian Network Classifiers. Mach. Learn. 1997, 29, 131–163. [Google Scholar] [CrossRef] [Green Version]
  3. Grossman, D.; Domingos, P. Learning Bayesian Network classifiers by maximizing conditional likelihood. In Proceedings of the Twenty-First International Conference on Machine Learning, ICML 2004, Banff, AB, Canada, 4–8 July 2004; pp. 361–368. [Google Scholar]
  4. Greiner, R.; Zhou, W. Structural Extension to Logistic Regression: Discriminative Parameter Learning of Belief Net Classifiers. In Proceedings of the Eighteenth National Conference on Artificial Intelligence, Edmonton, AB, Canada, 28 July–1 August 2002; pp. 167–173. [Google Scholar]
  5. Lucas, P.J.F. Restricted Bayesian Network Structure Learning. In Proceedings of the First European Workshop on Probabilistic Graphical Models, Cuenca, Spain, 6–8 November 2002. [Google Scholar]
  6. Carvalho, A.M.; Adão, P.; Mateus, P. Efficient Approximation of the Conditional Relative Entropy with Applications to Discriminative Learning of Bayesian Network Classifiers. Entropy 2013, 15, 2716–2735. [Google Scholar] [CrossRef] [Green Version]
  7. Mihaljević, B.; Bielza, C.; Larrañaga, P. Learning Bayesian network classifiers with completed partially directed acyclic graphs. In Proceedings of the Ninth International Conference on Probabilistic Graphical Models, Prague, Czech Republic, 11–14 September 2018; Volume 72, pp. 272–283. [Google Scholar]
  8. Koivisto, M.; Sood, K. Exact Bayesian Structure Discovery in Bayesian Networks. J. Mach. Learn. Res. 2004, 5, 549–573. [Google Scholar]
  9. Singh, A.P.; Moore, A.W. Finding Optimal Bayesian Networks by Dynamic Programming; Technical Report; Carnegie Mellon University: Pittsburgh, PA, USA, 2005. [Google Scholar]
  10. Silander, T.; Myllymäki, P. A Simple Approach for Finding the Globally Optimal Bayesian Network Structure. In Proceedings of the Uncertainty in Artificial Intelligence, Cambridge, MA, USA, 13–16 July 2006; pp. 445–452. [Google Scholar]
  11. De Campos, C.P.; Ji, Q. Efficient Structure Learning of Bayesian Networks Using Constraints. J. Mach. Learn. Res. 2011, 12, 663–689. [Google Scholar]
  12. Malone, B.M.; Yuan, C.; Hansen, E.A.; Bridges, S. Improving the Scalability of Optimal Bayesian Network Learning with External-Memory Frontier Breadth-First Branch and Bound Search. In Proceedings of the Uncertainty in Artificial Intelligence, Barcelona, Spain, 14–17 July 2011. [Google Scholar]
  13. Yuan, C.; Malone, B. Learning Optimal Bayesian Networks: A Shortest Path Perspective. J. Artif. Intell. Res. 2013, 48, 23–65. [Google Scholar] [CrossRef]
  14. Cussens, J. Bayesian network learning with cutting planes. In Proceedings of the 27th Conference on Uncertainty in Artificial Intelligence, Catalina Island, CA, USA, 14–18 August 2012; pp. 153–160. [Google Scholar]
  15. Barlett, M.; Cussens, J. Advances in Bayesian Network Learning Using Integer Programming. In Proceedings of the Twenty-Ninth Conference on Uncertainty in Artificial Intelligence, Bellevue, WA, USA, 11–15 August 2013; pp. 182–191. [Google Scholar]
  16. Suzuki, J. A theoretical analysis of the BDeu scores in Bayesian network structure learning. Behaviormetrika 2017, 44, 97–116. [Google Scholar] [CrossRef] [Green Version]
  17. Koller, D.; Friedman, N. Probabilistic Graphical Models: Principles and Techniques; MIT Press: Cambridge, MA, USA, 2009. [Google Scholar]
  18. Verma, T.; Pearl, J. Equivalence and Synthesis of Causal Models. In Proceedings of the Sixth Annual Conference on Uncertainty in Artificial Intelligence (UAI ’90), Cambridge, MA, USA, 27–29 July 1990; Elsevier Science Inc.: Amsterdam, The Netherlands, 1990; pp. 255–270. [Google Scholar]
  19. Chickering, D.M. Learning Equivalence Classes of Bayesian-network Structures. J. Mach. Learn. Res. 2002, 2, 445–498. [Google Scholar] [CrossRef]
  20. Heckerman, D.; Geiger, D.; Chickering, D.M. Learning Bayesian Networks: The Combination of Knowledge and Statistical Data. Mach. Learn. 1995, 20, 197–243. [Google Scholar] [CrossRef] [Green Version]
  21. Buntine, W. Theory Refinement on Bayesian Networks. In Proceedings of the Seventh Conference on Uncertainty in Artificial Intelligence, Los Angeles, CA, USA, 13–15 July 1991; Morgan Kaufmann Publishers Inc.: San Francisco, CA, USA, 1991; pp. 52–60. [Google Scholar]
  22. Silander, T.; Kontkanen, P.; Myllymäki, P. On Sensitivity of the MAP Bayesian Network Structure to the Equivalent Sample Size Parameter. In Proceedings of the Twenty-Third Conference on Uncertainty in Artificial Intelligence (UAI’07), Vancouver, BC, Canada, 19–22 July 2007; pp. 360–367. [Google Scholar]
  23. Steck, H. Learning the Bayesian Network Structure: Dirichlet Prior vs. Data. In Proceedings of the 24th Conference in Uncertainty in Artificial Intelligence (UAI 2008), Helsinki, Finland, 9–12 July 2008; pp. 511–518. [Google Scholar]
  24. Ueno, M. Learning likelihood-equivalence Bayesian networks using an empirical Bayesian approach. Behaviormetrika 2008, 35, 115–135. [Google Scholar] [CrossRef]
  25. Ueno, M. Learning Networks Determined by the Ratio of Prior and Data. In Proceedings of the Uncertainty in Artificial Intelligence, Catalina Island, CA, USA, 8–11 July 2010; pp. 598–605. [Google Scholar]
  26. Ueno, M. Robust learning Bayesian networks for prior belief. In Proceedings of the Uncertainty in Artificial Intelligence, Barcelona, Spain, 14–17 July 2011; pp. 689–707. [Google Scholar]
  27. Lichman, M. UCI Machine Learning Repository; School of Information and Computer Science, University of California: Irvine, CA, USA, 2013. [Google Scholar]
  28. De Campos, C.P.; Cuccu, M.; Corani, G.; Zaffalon, M. Extended Tree Augmented Naive Classifier. In Proceedings of the 7th European Workshop on Probabilistic Graphical Models, Utrecht, The Netherlands, 17–19 September 2014; van der Gaag, L.C., Feelders, A.J., Eds.; Springer International Publishing: Cham, Switzerland, 2014; pp. 176–189. [Google Scholar]
  29. Acid, S.; de Campos, L.M.; Castellano, J.G. Learning Bayesian Network Classifiers: Searching in a Space of Partially Directed Acyclic Graphs. Mach. Learn. 2005, 59, 213–235. [Google Scholar] [CrossRef] [Green Version]
  30. Tsamardinos, I.; Brown, L.E.; Aliferis, C.F. The Max-min Hill-climbing Bayesian Network Structure Learning Algorithm. Mach. Learn. 2006, 65, 31–78. [Google Scholar] [CrossRef] [Green Version]
  31. Scutari, M. Learning Bayesian Networks with the bnlearn R Package. J. Stat. Softw. Artic. 2010, 35, 1–22. [Google Scholar] [CrossRef] [Green Version]
  32. Niinimäki, T.; Parviainen, P. Local Structure Discovery in Bayesian Networks. In Proceedings of the Twenty-Eighth Conference on Uncertainty in Artificial Intelligence (UAI’12), Catalina Island, CA, USA, 14–18 August 2012; pp. 634–643. [Google Scholar]
  33. Gao, T.; Ji, Q. Efficient score-based Markov Blanket discovery. Int. J. Approx. Reason. 2017, 80, 277–293. [Google Scholar] [CrossRef]
  34. Aliferis, C.F.; Tsamardinos, I.; Statnikov, A. HITON: A Novel Markov Blanket Algorithm for Optimal Variable Selection. In AMIA Annual Symposium Proceedings; American Medical Informatics Association: Bethesda, MD, USA, 2003; pp. 21–25. [Google Scholar]
  35. Towards scalable and data efficient learning of Markov boundaries. Int. J. Approx. Reason. 2007, 45, 211–232. [CrossRef] [Green Version]
  36. Sullivan, G.M.; Feinn, R. Using Effect Size—Or Why the p Value Is Not Enough. J. Grad. Med Educ. 2012, 4, 279–282. [Google Scholar] [CrossRef] [Green Version]
  37. Cover, T.M.; Thomas, J.A. Elements of Information Theory; Wiley-Interscience: Hoboken, NJ, USA, 1991. [Google Scholar]
  38. Steck, H.; Jaakkola, T.S. On the Dirichlet Prior and Bayesian Regularization. In Proceedings of the 15th International Conference on Neural Information Processing Systems (NIPS’02), Vancouver, BC, Canada, 9–14 December 2002; MIT Press: Cambridge, MA, USA, 2002; pp. 713–720. [Google Scholar]
  39. Kass, R.E.; Raftery, A.E. Bayes Factors. J. Am. Stat. Assoc. 1995, 90, 773–795. [Google Scholar] [CrossRef]
  40. Natori, K.; Uto, M.; Nishiyama, Y.; Kawano, S.; Ueno, M. Constraint-Based Learning Bayesian Networks Using Bayes Factor. In Proceedings of the Second International Workshop on Advanced Methodologies for Bayesian Networks, Yokohama, Japan, 16–18 November 2015; Volume 9505, pp. 15–31. [Google Scholar]
  41. Natori, K.; Uto, M.; Ueno, M. Consistent Learning Bayesian Networks with Thousands of Variables. In Proceedings of the Machine Learning Research, Kyoto, Japan, 20–22 September 2017; Volume 73, pp. 57–68. [Google Scholar]
  42. Hommel, G. A Stagewise Rejective Multiple Test Procedure Based on a Modified Bonferroni Test. Biometrika 1988, 75, 383–386. [Google Scholar] [CrossRef]
  43. Demšar, J. Statistical Comparisons of Classifiers over Multiple Data Sets. J. Mach. Learn. Res. 2006, 7, 1–30. [Google Scholar]
  44. Ling, C.X.; Zhang, H. The Representational Power of Discrete Bayesian Networks. J. Mach. Learn. Res. 2003, 3, 709–721. [Google Scholar]
  45. Choi, A.; Wang, R.; Darwiche, A. On the relative expressiveness of Bayesian and neural networks. Int. J. Approx. Reason. 2019, 113, 303–323. [Google Scholar] [CrossRef] [Green Version]
  46. Shih, A.; Choi, A.; Darwiche, A. A Symbolic Approach to Explaining Bayesian Network Classifiers. In Proceedings of the Twenty-Seventh International Joint Conference on Artificial Intelligence, IJCAI-18, Stockholm, Sweden, 13–19 July 2018; pp. 5103–5111. [Google Scholar] [CrossRef] [Green Version]
  47. Shih, A.; Choi, A.; Darwiche, A. Compiling Bayesian Network Classifiers into Decision Graphs. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019; pp. 7966–7974. [Google Scholar]
  48. Darwiche, A.; Hirth, A. On The Reasons Behind Decisions. arXiv 2020, arXiv:2002.09284. [Google Scholar]
  49. Darwiche, A. Three Modern Roles for Logic in AI. In Proceedings of the 39th ACM SIGMOD-SIGACT-SIGAI Symposium on Principles of Database Systems, Portland, OR, USA, 14–19 June 2020; pp. 229–243. [Google Scholar] [CrossRef]
  50. Sugahara, S.; Aomi, I.; Ueno, M. Bayesian Network Model Averaging Classifiers by Subbagging. In Proceedings of the 10th International Conference on Probabilistic Graphical Models, Skørping, Denmark, 23–25 September 2020. [Google Scholar]
  51. Sugahara, S.; Uto, M.; Ueno, M. Exact learning augmented naive Bayes classifier. In Proceedings of the 9th International Conference on Probabilistic Graphical Models, Prague, Czech Republic, 11–14 September 2018; Volume 72, pp. 439–450. [Google Scholar]
Figure 1. A network which satisfies Assumptions 2 and 3 (CANCER network [31]).
Figure 1. A network which satisfies Assumptions 2 and 3 (CANCER network [31]).
Entropy 23 01703 g001
Figure 2. A network which violates Assumptions 2 and 3 (ASIA network [31]).
Figure 2. A network which violates Assumptions 2 and 3 (ASIA network [31]).
Entropy 23 01703 g002
Table 1. Seven methods compared in the experiments.
Table 1. Seven methods compared in the experiments.
AbbreviationMethods
Naive BayesNavie Bayes classifier.
GBN-BDeuExact learning GBN method by maximizing BDeu.
GBN-CMDL [3]Greedy learning GBN method using the hill-climbing search by minimizing CMDL while estimating parameters by maximizing LL.
BNC2P [3]Greedy learning method with at most two parents per variable using the hill-climbing search by maximizing CLL while estimating parameters by maximizing LL.
TAN-aCLL [6]Exact learning TAN method by maximizing aCLL.
gGBN-BDeuGreedy learning GBN method using hill-climbing by maximizing BDeu.
MC-DAGGES [7]Greedy learning method in the space of the Markov equivalent classes of MC-DAGs using the greedy equivalence search [19] by maximizing CLL while estimating parameters by maximizing LL.
Table 2. Computational environment.
Table 2. Computational environment.
CPU2.2 GHz XEON 10-core processor
System Memory128 GB
OSWindows 10
SoftwareJava
Table 3. Classification accuracies of GBN-BDeu, ANB-BDeu, fsANB-BDeu, and traditional methods (bold text signifies the highest accuracy).
Table 3. Classification accuracies of GBN-BDeu, ANB-BDeu, fsANB-BDeu, and traditional methods (bold text signifies the highest accuracy).
No.DatasetVariablesSample SizeClassesNaive BayesGBN-CMDLBNC2PTAN-aCLLgGBN-BDeuMC-DAGGESGBN-BDeuANB-BDeufsANB-BDeu
1Balance Scale536250.91520.33330.85600.86560.91520.74320.91520.91520.9152
2banknote authentication5213720.84330.88190.87970.87610.88190.87680.88120.88120.8812
3Hayes–Roth531320.81820.61360.68940.67420.75250.69700.61360.81820.8333
4iris531500.71330.78000.82000.82000.81330.78000.82670.82000.8200
5lenses53240.75000.83330.66670.70830.83330.83330.83330.75000.8750
6Car Evaluation7417280.85710.94970.94160.94330.94160.91260.94160.94270.9416
7liver723450.63190.61450.62900.66090.60290.64350.60870.63480.6377
8MONK’s Problems724320.75001.00001.00001.00000.84491.00001.00001.00001.0000
9mux672640.54690.37500.56250.46880.40630.76560.45310.54690.5547
10LED781032000.72940.73660.73750.73500.72970.73310.72940.72940.7294
11HTRU29217,8980.70310.70960.70700.70180.71880.72140.73050.71880.7161
12Nursery9512,9600.67820.71260.60920.58620.71260.63220.71260.67820.7126
13pima927680.89660.90860.91180.91300.90920.90930.91120.91410.9141
14post93870.90330.58230.94420.91770.92910.90460.93400.91810.9177
15Breast Cancer1022770.97510.89170.94730.94880.70580.63540.97510.97510.9751
16Breast Cancer Wisconsin1026830.74010.62090.68230.71840.70940.97800.71840.70400.7473
17Contraceptive Method Choice10314730.46710.45010.47450.47050.44400.45760.45420.46500.4725
18glass1062140.55610.56540.57940.63080.46260.58880.57010.64490.5888
19shuttle-small10658000.93840.96600.97030.95830.96830.95860.96930.97160.9695
20threeOf91025120.81640.94340.86910.88280.86520.87500.88870.87300.8633
21Tic-Tac-Toe1029580.69210.88410.73380.72030.67540.75570.83400.84970.8570
22MAGIC Gamma Telescope11219,0200.74820.78490.78060.76310.78440.77810.78730.78740.7865
23Solar Flare11913890.78110.82650.83150.82290.84310.80130.84310.82290.8373
24heart1422700.82590.81850.80370.81480.82220.83330.82590.81850.8296
25wine1431780.92700.94380.91570.93260.90450.94380.92700.92700.9270
26cleve1422960.84120.82090.80070.83780.79730.80410.79730.82770.8243
27Australian1526900.82900.83120.83480.84640.84200.84060.85360.82460.8522
28crx1526530.83770.83460.82080.85600.86220.85760.85910.85150.8591
29EEG15214,9800.57780.67870.63740.61250.67320.61820.68140.68640.6864
30Congressional Voting Records1722320.90950.96980.96120.91810.97410.90090.96550.94830.9397
31zoo1751010.98020.91090.95051.00000.95050.98020.93070.95050.9604
32pendigits171010,9920.80320.90620.87190.87000.92530.83590.92900.92790.9279
33letter172620,0000.44660.57960.51320.50930.57610.46640.57610.59350.5881
34ClimateModel1925400.92220.94070.92410.93330.93700.92960.90000.84260.9278
35Image Segmentation19723100.72900.79180.79910.74070.80260.74760.81560.82250.8225
36lymphography1941480.84460.79390.79730.83110.79050.80410.75000.77700.7838
37vehicle1948460.43500.59100.59100.58160.54610.54140.57680.62530.6217
38hepatitis202800.85000.73750.88750.87500.85000.88750.58750.62500.8375
39German21210000.74300.61100.73400.74700.71400.71800.72100.73800.7410
40bank21230,4880.85440.86180.89280.86180.89520.87080.89560.89500.8953
41waveform-2122350000.78860.78620.77540.78960.76980.79260.78460.79660.7972
42Mushroom22256440.99571.00001.00000.99951.00000.99860.99491.00001.0000
43spect2322630.79400.79400.79030.80900.76030.80520.73780.82400.8240
average 0.77640.77210.79360.79430.78670.79440.79630.80610.8184
p-value (ANB-BDeu vs. the other methods) 0.003080.041360.006720.056140.068760.060100.22628--
p-value (fsANB-BDeu vs. the other methods) 0.000010.000140.000130.002800.000150.002120.000640.01101-
Table 4. Statistical summary of GBN-BDeu and fsANB-BDeu.
Table 4. Statistical summary of GBN-BDeu and fsANB-BDeu.
No.VariablesClassesSample SizeParentsChildrenSparse DataMB SizeMax ParentsRemoved Variables
1536250.43.60.04.01.00.0
25213720.02.00.04.04.00.0
3531323.00.017.23.01.01.0
4531501.81.20.03.02.00.0
553241.11.00.02.11.12.0
67417282.03.00.05.02.01.0
7723450.01.90.03.42.00.1
8724323.00.00.03.03.00.0
972645.80.05.25.81.02.1
1081032000.96.10.07.01.00.0
119217,8981.84.20.04.22.00.9
129512,9604.03.00.00.00.08.0
13927681.41.70.07.04.00.0
1493870.00.00.07.03.00.1
151022770.98.00.01.01.00.0
161026830.70.30.08.92.05.0
1710314730.70.80.01.72.50.6
181062140.63.10.04.32.72.0
1910658002.04.00.07.05.01.9
201025125.02.10.07.62.70.2
211029581.22.20.05.33.00.3
2211219,0200.06.10.08.04.01.7
2311913890.80.20.01.02.05.3
241422701.84.20.06.32.01.8
251431781.75.30.08.12.10.0
261422961.84.50.06.62.03.1
271526901.42.80.04.52.83.3
281526531.32.80.04.22.22.7
2915214,9800.48.20.012.85.00.0
301722321.32.60.16.23.81.8
311751014.31.620.37.45.11.2
32171010,9922.613.40.116.05.60.0
33172620,0002.99.10.013.05.02.0
341925401.84.40.016.61.012.9
3519723100.710.40.013.24.00.0
361941481.65.90.213.12.25.3
371948461.15.10.110.14.10.5
38202801.36.10.416.06.95.4
3921210001.12.80.04.12.17.4
4021230,4884.12.032.59.96.04.0
4122350003.810.10.014.54.02.0
4222256441.33.39.06.46.40.0
432322632.03.40.07.73.00.0
Table 5. The SHD between the structure learned by the proposed method and the I-map with the fewest parameters among all the ANB structures and the KLD between the learned class variable posterior by the proposed method and learned one using the true structure.
Table 5. The SHD between the structure learned by the proposed method and the I-map with the fewest parameters among all the ANB structures and the KLD between the learned class variable posterior by the proposed method and learned one using the true structure.
NetworkVariablesSample SizeSHD-(Proposal, I-Map ANB)KLD-(Proposal, True Structure)
1003 2.31 × 10 2
5002 1.24 × 10 1
10002 7.63 × 10 2
ASIA850001 3.67 × 10 3
10,0000 9.26 × 10 4
50,0000 6.28 × 10 4
100,0000 3.59 × 10 5
1001 8.79 × 10 2
5001 2.43 × 10 3
100000.00
CANCER5500000.00
10,00000.00
50,00000.00
100,00000.00
Table 6. Missing variables, extra variables, and runtimes (ms) of each method.
Table 6. Missing variables, extra variables, and runtimes (ms) of each method.
NetworkVariablesMMPCHITON-PCPCMB S 2 TMBProposal
MissingExtraRuntimeMissingExtraRuntimeMissingExtraRuntimeMissingExtraRuntimeMissingExtraRuntime
ASIA81.250.002511.750.631171.750.631630.250.508880.003.5013
SACHS111.910.0010622.640.362482.000.006100.000.0048420.002.5512
CHILD201.750.0567562.350.953802.000.2511910.050.0566690.0011.8016
WATER323.590.004074.000.191403.780.312602.031.4729,5270.2513.4425
ALARM371.810.1438322.380.572812.190.1910250.140.1111,2720.0510.9239
BARLEY482.851.2349283.460.422693.190.428301.150.4699,2900.389.7549
Average 2.190.2428722.760.522392.480.306800.600.4325,4150.118.6626
Table 7. Average classification accuracy of each method.
Table 7. Average classification accuracy of each method.
MMPCHITON-PCPCMB S 2 TMBProposal
Average0.61850.62190.63020.79800.8164
Table 8. Runtimes (ms) of GBN-BDeu, fsANB-BDeu, and the proposed PC search method.
Table 8. Runtimes (ms) of GBN-BDeu, fsANB-BDeu, and the proposed PC search method.
No.VariablesSample SizeClassesGBN-BDeufsANB-BDeuThe Proposed PC Search Method
156253169.423.06.3
251372219.310.32.0
35132315.63.00.2
45150316.75.00.2
5524315.31.00.1
671728490.822.91.7
77345221.115.60.3
87432231.020.70.5
9764218.99.10.1
108320010114.655.13.1
11917,8982300.5251.310.2
12912,9603707.4525.85.8
139768966.827.60.6
14987539.60.30.1
15102772162.66.90.3
16106832453.1258.90.4
171014733161.1121.40.8
1810214663.022.30.2
191058006159.667.22.8
20105122102.758.20.4
21109582212.2193.00.5
221119,0202979.8277.25.3
231113899379.417.20.9
241427021988.6299.80.1
251417831233.7585.00.1
261429622034.5115.20.2
2715690210,700.3927.60.3
2815653223,069.52774.30.2
291514,980212,407.68248.84.1
3017232211,682.61623.60.2
311710157326.51985.10.1
321710,9921084,967.148,636.93.4
331720,00026339,910.230,224.86.3
34195402217,457.012.00.3
351923107190,895.9103,447.51.0
36191484107,641.81171.40.2
37198464144,669.562,663.00.4
382080298,841.9821.60.1
3921100022,706,616.68885.10.5
402130,48821,562,6734.5130,491.611.8
41225000310,022,030.7757,611.72.1
4222564424,640,293.52,382,657.72.3
432326322,553,290.41,386,088.20.2
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Sugahara, S.; Ueno, M. Exact Learning Augmented Naive Bayes Classifier. Entropy 2021, 23, 1703. https://0-doi-org.brum.beds.ac.uk/10.3390/e23121703

AMA Style

Sugahara S, Ueno M. Exact Learning Augmented Naive Bayes Classifier. Entropy. 2021; 23(12):1703. https://0-doi-org.brum.beds.ac.uk/10.3390/e23121703

Chicago/Turabian Style

Sugahara, Shouta, and Maomi Ueno. 2021. "Exact Learning Augmented Naive Bayes Classifier" Entropy 23, no. 12: 1703. https://0-doi-org.brum.beds.ac.uk/10.3390/e23121703

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop