Please use this identifier to cite or link to this item:
Document Type: Artigo
Title: Parallel implementation of Expectation-Maximisation algorithm for the training of Gaussian Mixture Models
Authors: Araújo, Gabriel Ferreira
Macedo, Hendrik Teixeira
Chella, Marco Túlio
Estombelo Montesco, Carlos Alberto
Medeiros, Marcus Vinícius Oliveira
Issue Date: Jul-2014
Abstract: Most machine learning algorithms need to handle large data sets. This feature often leads to limitations on processing time and memory. The Expectation-Maximization (EM) is one of such algorithms, which is used to train one of the most commonly used parametric statistical models, the Gaussian Mixture Models (GMM). All steps of the algorithm are potentially parallelizable once they iterate over the entire data set. In this study, we propose a parallel implementation of EM for training GMM using CUDA. Experiments are performed with a UCI dataset and results show a speedup of 7 if compared to the sequential version. We have also carried out modifications to the code in order to provide better access to global memory and shared memory usage. We have achieved up to 56.4% of achieved occupancy, regardless the number of Gaussians considered in the set of experiments.
Keywords: Expectation-Maximization (EM)
Gaussian Mixture Models (GMM)
Modelo de misturas guassianas
ISSN: 1552-6607
Publisher / Institution : Science Publications
Citation: ARAÚJO, G. F. et al. Parallel implementation of Expectation-Maximisation algorithm for the training of Gaussian Mixture Models. Journal of Computer Science, v. 10, n. 10, jul. 2014. Disponível em: <>. Acesso em: 16 maio 2016.
License: Creative Commons Attribution License
Appears in Collections:DCOMP - Artigos de periódicos

Files in This Item:
File Description SizeFormat 
ExpectationMaximisationAlgorithm.pdf226,02 kBAdobe PDFThumbnail

Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.