4.5 Article

Stochastic optimization for bayesian network classifiers

Journal

APPLIED INTELLIGENCE
Volume 52, Issue 13, Pages 15496-15516

Publisher

SPRINGER
DOI: 10.1007/s10489-022-03356-z

Keywords

Bayesian network classifiers; Ensemble learning; Stochastic optimization; Random sampling

Funding

  1. NationalKeyResearch and Development Program of China [2019YFC1804804]
  2. Open Research Project of The Hubei Key Laboratory of Intelligent Geo-Information Processing
  3. Scientific and Technological Developing Scheme of Jilin Province [20200201281JC]

Ask authors/readers for more resources

This paper addresses two important issues in learning Bayesian network classifiers: reducing the complexity of network topology and making the learned joint probability distribution fit the data. Ensemble learning algorithms are used to achieve the tradeoff between bias and variance by transforming high-order topology into low-order ones. The proposed algorithm, called random Bayesian forest (RBF), achieves remarkable classification performance compared to state-of-the-art out-of-core BNCs.
How to reduce the complexity of network topology and make the learned joint probability distribution fit data are two important but inconsistent issues for learning Bayesian network classifier (BNC). By transforming one single high-order topology into a set of low-order ones, ensemble learning algorithms can include more hypothesis implicated in training data and help achieve the tradeoff between bias and variance. Resampling from training data can vary the results of member classifiers of the ensemble, whereas the potentially lost information may bias the estimate of conditional probability distribution and then introduce insignificant rather than significant dependency relationships into the network topology of BNC. In this paper, we propose to learn from training data as a whole and apply heuristic search strategy to flexibly identify the significant conditional dependencies, and then the attribute order is determined implicitly. Random sampling is introduced to make each member of the ensemble unstable and fully represent the conditional dependencies. The experimental evaluation on 40 UCI datasets reveals that the proposed algorithm, called random Bayesian forest (RBF), achieves remarkable classification performance compared to the extended version of state-of-the-art out-of-core BNCs (e.g., SKDB, WATAN, WAODE, SA2DE, SASA2DE and IWAODE).

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.5
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available