File Download
  Links for fulltext
     (May Require Subscription)
Supplementary

Article: An imputation-regularized optimization algorithm for high dimensional missing data problems and beyond

TitleAn imputation-regularized optimization algorithm for high dimensional missing data problems and beyond
Authors
KeywordsExpectation–maximization algorithm
Gaussian graphical model
Gibbs sampler
Imputation consistency
Random‐coefficient model
Issue Date2018
PublisherWiley-Blackwell Publishing Ltd. The Journal's web site is located at http://www.blackwellpublishing.com/journals/RSSB
Citation
Journal of the Royal Statistical Society. Series B: Statistical Methodology, 2018, v. 80 n. 5, p. 899-926 How to Cite?
AbstractMissing data are frequently encountered in high-dimensional problems, but they are usually difficult to deal with using standard algorithms, such as the expectation-maximization (EM) algorithm and its variants. To tackle this difficulty, some problem-specific algorithms have been developed in the literature, but there still lacks a general algorithm. This work is to fill the gap: we propose a general algorithm for high-dimensional missing data problems. The proposed algorithm works by iterating between an imputation step and a consistency step. At the imputation step, the missing data are imputed conditional on the observed data and the current estimate of parameters; and at the consistency step, a consistent estimate is found for the minimizer of a Kullback-Leibler divergence defined on the pseudo-complete data. For high dimensional problems, the consistent estimate can be found under sparsity constraints. The consistency of the averaged estimate for the true parameter can be established under quite general conditions. The proposed algorithm is illustrated using high-dimensional Gaussian graphical models, high-dimensional variable selection, and a random coefficient model.
Persistent Identifierhttp://hdl.handle.net/10722/272098
ISSN
2023 Impact Factor: 3.1
2023 SCImago Journal Rankings: 4.330
ISI Accession Number ID

 

DC FieldValueLanguage
dc.contributor.authorLuo, Y-
dc.contributor.authorLiang, FAMING-
dc.contributor.authorJia, BOCHAO-
dc.contributor.authorXue, JINGNAN-
dc.contributor.authorLi, QIZAI-
dc.date.accessioned2019-07-20T10:35:38Z-
dc.date.available2019-07-20T10:35:38Z-
dc.date.issued2018-
dc.identifier.citationJournal of the Royal Statistical Society. Series B: Statistical Methodology, 2018, v. 80 n. 5, p. 899-926-
dc.identifier.issn1369-7412-
dc.identifier.urihttp://hdl.handle.net/10722/272098-
dc.description.abstractMissing data are frequently encountered in high-dimensional problems, but they are usually difficult to deal with using standard algorithms, such as the expectation-maximization (EM) algorithm and its variants. To tackle this difficulty, some problem-specific algorithms have been developed in the literature, but there still lacks a general algorithm. This work is to fill the gap: we propose a general algorithm for high-dimensional missing data problems. The proposed algorithm works by iterating between an imputation step and a consistency step. At the imputation step, the missing data are imputed conditional on the observed data and the current estimate of parameters; and at the consistency step, a consistent estimate is found for the minimizer of a Kullback-Leibler divergence defined on the pseudo-complete data. For high dimensional problems, the consistent estimate can be found under sparsity constraints. The consistency of the averaged estimate for the true parameter can be established under quite general conditions. The proposed algorithm is illustrated using high-dimensional Gaussian graphical models, high-dimensional variable selection, and a random coefficient model.-
dc.languageeng-
dc.publisherWiley-Blackwell Publishing Ltd. The Journal's web site is located at http://www.blackwellpublishing.com/journals/RSSB-
dc.relation.ispartofJournal of the Royal Statistical Society. Series B: Statistical Methodology-
dc.rightsPreprint This is the pre-peer reviewed version of the following article: [Journal of the Royal Statistical Society. Series B: Statistical Methodology, 2018, v. 80 n. 5, p. 899-926], which has been published in final form at [http://dx.doi.org/10.1111/rssb.12279]. This article may be used for non-commercial purposes in accordance with Wiley Terms and Conditions for Use of Self-Archived Versions.-
dc.subjectExpectation–maximization algorithm-
dc.subjectGaussian graphical model-
dc.subjectGibbs sampler-
dc.subjectImputation consistency-
dc.subjectRandom‐coefficient model-
dc.titleAn imputation-regularized optimization algorithm for high dimensional missing data problems and beyond-
dc.typeArticle-
dc.identifier.emailLuo, Y: kurtluo@hku.hk-
dc.identifier.authorityLuo, Y=rp02428-
dc.description.naturepreprint-
dc.identifier.doi10.1111/rssb.12279-
dc.identifier.scopuseid_2-s2.0-85055455509-
dc.identifier.hkuros299561-
dc.identifier.volume80-
dc.identifier.issue5-
dc.identifier.spage899-
dc.identifier.epage926-
dc.identifier.isiWOS:000448897700003-
dc.publisher.placeUnited Kingdom-
dc.identifier.issnl1369-7412-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats