Loading...
Please wait, while we are loading the content...
Similar Documents
Scaling bayesian network parameter learning with expectation maximization using mapreduce.
| Content Provider | CiteSeerX |
|---|---|
| Author | Reed, Erik B. Mengshoel, Ole J. |
| Abstract | Bayesian network (BN) parameter learning from incomplete data can be a computationally expensive task for incomplete data. Applying the EM algorithm to learn BN parameters is unfortunately susceptible to local optima and prone to premature convergence. We develop and experiment with two methods for improving EM parameter learning by using MapReduce: Age-Layered Expectation Maximization (ALEM) and Multiple Expectation Maximization (MEM). Leveraging MapReduce for distributed machine learning, these algorithms (i) operate on a (potentially large) population of BNs and (ii) partition the data set as is traditionally done with MapReduce machine learning. For example, we achieved gains using the Hadoop implementation of MapReduce in both parameter quality (likelihood) and number of iterations (runtime) using distributed ALEM on for the BN Asia over 20,000 MEM and ALEM trials. 1 |
| File Format | |
| Access Restriction | Open |
| Subject Keyword | Bayesian Network Parameter Learning Expectation Maximization Incomplete Data Mapreduce Machine Learning Bayesian Network Alem Trial Bn Parameter Parameter Quality Expensive Task Hadoop Implementation Machine Learning Data Set Multiple Expectation Maximization Bn Asia Local Optimum Age-layered Expectation Maximization Em Algorithm Em Parameter Learning |
| Content Type | Text |