The POWER (PrObabilistic Weighted Ensemble of Roles) model[1] is a Bayesian mixture model where a single data point is generated by multiple components with varying degrees of influence. This is unlike standard mixture models where each data point is generated by a single mixture component. The POWER model allows for capturing various hidden and complexly overlapping components and roles of entities that contribute to generating the data points. However, the POWER model suffers from a very slow learning time. The highest complexity time of the parameters' learning steps is , with n being the number of data points, k the number of components of the model and d the number of data attributes. In this paper, we propose an approximation to the POWER model weight parameter learning algorithm that reduces the computational time significantly. The overall complexity of the new algorithm is in practice (where p is an approximation parameter much smaller than d). This allows the model learning time to be linear in the number of attributes of the data and provides a significant speedup over the original algorithm. We demonstrate the accuracy of our approximation technique using synthetic and real datasets. We also provide experimental results of the approximate POWER model on a dataset of NIPS papers and a dataset of wireless web access patterns and show that the model learnt are similar. An implementation of the approximate POWER model on the dataset of NIPS papers is about 27 times faster than the original version. © 2012 Springer-Verlag.
CITATION STYLE
Almutairi, A., Ranka, S., & Somaiya, M. (2012). A fast algorithm for learning weighted ensemble of roles. In Communications in Computer and Information Science (Vol. 306 CCIS, pp. 73–85). https://doi.org/10.1007/978-3-642-32129-0_13
Mendeley helps you to discover research relevant for your work.