Carnegie Mellon University
Browse
- No file added yet -

Scaling Bayesian Network Parameter Learning with Expectation Maximization using MapReduce

Download (309.82 kB)
journal contribution
posted on 2012-12-01, 00:00 authored by Erik B Reed, Ole J Mengshoel
Bayesian network (BN) parameter learning from incomplete data can be a computationally expensive task for incomplete data. Applying the EM algorithm to learn BN parameters is unfortunately susceptible to local optima and prone to premature convergence. We develop and experiment with two methods for improving EM parameter learning by using MapReduce: Age-Layered Expectation Maximization (ALEM) and Multiple Expectation Maximization (MEM). Leveraging MapReduce for distributed machine learning, these algorithms (i) operate on a (potentially large) population of BNs and (ii) partition the data set as is traditionally done with MapReduce machine learning. For example, we achieved gains using the Hadoop implementation of MapReduce in both parameter quality (likelihood) and number of iterations (runtime) using distributed ALEM for the BN Asia over 20,000 MEM and ALEM trials.

History

Date

2012-12-01

Usage metrics

    Exports

    RefWorks
    BibTeX
    Ref. manager
    Endnote
    DataCite
    NLM
    DC