Efficient parallel learning of hidden Markov chain models on SMPs

3Citations
Citations of this article
9Readers
Mendeley users who have this article in their library.

Abstract

Quad-core cpus have been a common desktop configuration for today's office. The increasing number of processors on a single chip opens new opportunity for parallel computing. Our goal is to make use of the multi-core as well as multi-processor architectures to speed up large-scale data mining algorithms. In this paper, we present a general parallel learning framework, Cut-And-Stitch, for training hiddenMarkov chain models. Particularly, we propose two model-specific variants, CAS-LDS for learning linear dynamical systems (LDS) and CAS-HMM for learning hiddenMarkov models (HMM). Our main contribution is a novel method to handle the data dependencies due to the chain structure of hidden variables, so as to parallelize the EM-based parameter learning algorithm. We implement CAS-LDS and CAS-HMM using OpenMP on two supercomputers and a quad-core commercial desktop. The experimental results show that parallel algorithms using Cut-And-Stitch achieve comparable accuracy and almost linear speedups over the traditional serial version. Copyright © 2010 The Institute of Electronics, Information and Communication Engineers.

Cite

CITATION STYLE

APA

Li, L., Fu, B., & Faloutsos, C. (2010). Efficient parallel learning of hidden Markov chain models on SMPs. IEICE Transactions on Information and Systems, E93-D(6), 1330–1342. https://doi.org/10.1587/transinf.E93.D.1330

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free