We consider a finite mixture of Gaussian regression models for high-dimensional data, where the number of covariates may be much larger than the sample size. We propose to estimate the unknown conditional mixture density by a maximum likelihood estimator, restricted on relevant variables selected by an ℓ1-penalized maximum likelihood estimator. We get an oracle inequality satisfied by this estimator with a Jensen-Kullback-Leibler type loss. Our oracle inequality is deduced from a general model selection theorem for maximum likelihood estimators on a random model subcollection. We can derive the penalty shape of the criterion, which depends on the complexity of the random model collection.
CITATION STYLE
Devijver, E. (2015). Finite mixture regression: A sparse variable selection by model selection for clustering. Electronic Journal of Statistics, 9(2), 2642–2674. https://doi.org/10.1214/15-EJS1082
Mendeley helps you to discover research relevant for your work.