Learning Bayesian Networks with Local Structure

  • Friedman N
  • Goldszmidt M
N/ACitations
Citations of this article
147Readers
Mendeley users who have this article in their library.
Get full text

Abstract

In this paper we examine a novel addition to the known methods for learning Bayesian networks from data that improves the quality of the learned networks. Our approach explicitly represents and learns the local structure in the conditional probability tables (CPTs), that quantify these networks. This increases the space of possible models, enabling the representation of CPTs with a variable number of parameters that depends on the learned local structures. The resulting learning procedure is capable of inducing models that better emulate the real complexity of the interactions present in the data. We describe the theoretical foundations and practical aspects of learning local structures, as well as an empirical evaluation of the proposed method. This evaluation indicates that learning curves characterizing the procedure that exploits the local structure converge faster than these of the standard procedure. Our results also show that networks learned with local structure tend to be more complex (in terms of arcs), yet require less parameters.

Cite

CITATION STYLE

APA

Friedman, N., & Goldszmidt, M. (1998). Learning Bayesian Networks with Local Structure. In Learning in Graphical Models (pp. 421–459). Springer Netherlands. https://doi.org/10.1007/978-94-011-5014-9_15

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free