Deakin University
Browse

Scalable learning of Bayesian network classifiers

Version 2 2024-06-05, 07:44
Version 1 2020-03-03, 13:45
journal contribution
posted on 2024-06-05, 07:44 authored by AM Martínez, GI Webb, S Chen, Nayyar ZaidiNayyar Zaidi
Ever increasing data quantity makes ever more urgent the need for highly scalable learners that have good classification performance. Therefore, an out-of-core learner with excellent time and space complexity, along with high expressivity (that is, capacity to learn very complex multivariate probability distributions) is extremely desirable. This paper presents such a learner. We propose an extension to the k-dependence Bayesian classifier (KDB) that discriminatively selects a sub-model of a full KDB classifier. It requires only one additional pass through the training data, making it a three-pass learner. Our extensive experimental evaluation on 16 large data sets reveals that this out-of-core algorithm achieves competitive classification performance, and substantially better training and classification time than state-of-the-art in-core learners such as random forest and linear and non-linear logistic regression.

History

Journal

Journal of Machine Learning Research

Volume

17

Article number

ARTN 44

Location

Cambridge, Mass.

ISSN

1532-4435

eISSN

1533-7928

Language

English

Publication classification

C1.1 Refereed article in a scholarly journal

Publisher

MICROTOME PUBL