Data-Adaptive Shrinkage via the Hyperpenalized EM Algorithm

1Citations
Citations of this article
3Readers
Mendeley users who have this article in their library.
Get full text

Abstract

We propose an extension of the expectation–maximization (EM) algorithm, called the hyperpenalized EM (HEM) algorithm, that maximizes a penalized log-likelihood, for which some data are missing or unavailable, using a data-adaptive estimate of the penalty parameter. This is potentially useful in applications for which the analyst is unable or unwilling to choose a single value of a penalty parameter but instead can posit a plausible range of values. The HEM algorithm is conceptually straightforward and also very effective, and we demonstrate its utility in the analysis of a genomic data set. Gene expression measurements and clinical covariates were used to predict survival time. However, many survival times are censored, and some observations only contain expression measurements derived from a different assay, which together constitute a difficult missing data problem. It is desired to shrink the genomic contribution in a data-adaptive way. The HEM algorithm successfully handles both the missing data and shrinkage aspects of the problem.

Cite

CITATION STYLE

APA

Boonstra, P. S., Taylor, J. M. G., & Mukherjee, B. (2015). Data-Adaptive Shrinkage via the Hyperpenalized EM Algorithm. Statistics in Biosciences, 7(2), 417–431. https://doi.org/10.1007/s12561-015-9132-x

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free