4.7 Article

Hierarchical Independence Thresholding for learning Bayesian network classifiers

Journal

KNOWLEDGE-BASED SYSTEMS
Volume 212, Issue -, Pages -

Publisher

ELSEVIER
DOI: 10.1016/j.knosys.2020.106627

Keywords

Bayesian network; Hierarchical independence thresholding; Informational independence; Probabilistic independence; Adaptive thresholding

Funding

  1. National Science Foundation of China [61872164]
  2. Scientific and Technological Developing Scheme of Jilin Province, China [20200201281JC]

Ask authors/readers for more resources

Researchers proposed a novel framework called Hierarchical Independence Thresholding (HIT) for efficient identification of informational conditional independence and probabilistic conditional independence, which improves the fit of learned topology to the data. Experimental results demonstrate that applying HIT to BNCs can achieve competitive classification performance.
Bayesian networks are powerful tools for knowledge representation and inference under conditions of uncertainty. However, learning an optimal Bayesian network classifier (BNC) is an NP-hard problem since its topology complexity increases exponentially with the number of attributes. Researchers proposed to apply information-theoretic criteria to measure conditional dependence, and independence assumptions are introduced implicitly or explicitly to simplify the network topology of BNC. In this paper, we clarify the mapping relationship between conditional mutual information and local topology, and then illustrate that informational independence does not correspond to probabilistic independence, the criterion of probabilistic independence does not necessarily hold for the independence topology. A novel framework of semi-naive Bayesian operation, called Hierarchical Independence Thresholding (HIT), is presented to efficiently identify informational conditional independence and probabilistic conditional independence by applying an adaptive thresholding method, redundant edges will be filtered out and the learned topology will fit the data better. Extensive experimental evaluation on 58 publicly available datasets reveals that when HIT is applied to BNCs (such as tree augmented Naive Bayes or k-dependence Bayesian classifier), the final BNCs achieve competitive classification performance compared to state-of-the-art learners such as Random Forest and Logistic regression. (C) 2020 Elsevier B.V. All rights reserved.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.7
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available