Bayesian Network Model Averaging Classifiers by Subbagging

Autor: Shouta Sugahara, Itsuki Aomi, Maomi Ueno
Jazyk: angličtina
Rok vydání: 2022
Předmět:
Zdroj: Entropy, Vol 24, Iss 5, p 743 (2022)
Druh dokumentu: article
ISSN: 1099-4300
DOI: 10.3390/e24050743
Popis: When applied to classification problems, Bayesian networks are often used to infer a class variable when given feature variables. Earlier reports have described that the classification accuracy of Bayesian network structures achieved by maximizing the marginal likelihood (ML) is lower than that achieved by maximizing the conditional log likelihood (CLL) of a class variable given the feature variables. Nevertheless, because ML has asymptotic consistency, the performance of Bayesian network structures achieved by maximizing ML is not necessarily worse than that achieved by maximizing CLL for large data. However, the error of learning structures by maximizing the ML becomes much larger for small sample sizes. That large error degrades the classification accuracy. As a method to resolve this shortcoming, model averaging has been proposed to marginalize the class variable posterior over all structures. However, the posterior standard error of each structure in the model averaging becomes large as the sample size becomes small; it subsequently degrades the classification accuracy. The main idea of this study is to improve the classification accuracy using subbagging, which is modified bagging using random sampling without replacement, to reduce the posterior standard error of each structure in model averaging. Moreover, to guarantee asymptotic consistency, we use the K-best method with the ML score. The experimentally obtained results demonstrate that our proposed method provides more accurate classification than earlier BNC methods and the other state-of-the-art ensemble methods do.
Databáze: Directory of Open Access Journals
Nepřihlášeným uživatelům se plný text nezobrazuje