File Download
There are no files associated with this item.
Links for fulltext
(May Require Subscription)
- Publisher Website: 10.1214/15-AOS1364
- Scopus: eid_2-s2.0-85013178156
- PMID: 26783374
- WOS: WOS:000368022000008
- Find via
Supplementary
- Citations:
- Appears in Collections:
Article: Projected principal component analysis in factor models
Title | Projected principal component analysis in factor models |
---|---|
Authors | |
Keywords | High-dimensionality Semiparametric factor models Loading matrix modeling Sieve approximation |
Issue Date | 2016 |
Citation | Annals of Statistics, 2016, v. 44, n. 1, p. 219-254 How to Cite? |
Abstract | This paper introduces a Projected Principal Component Analysis (Projected-PCA), which employs principal component analysis to the projected (smoothed) data matrix onto a given linear space spanned by covariates. When it applies to high-dimensional factor analysis, the projection removes noise components. We show that the unobserved latent factors can be more accurately estimated than the conventional PCA if the projection is genuine, or more precisely, when the factor loading matrices are related to the projected linear space. When the dimensionality is large, the factors can be estimated accurately even when the sample size is finite. We propose a flexible semiparametric factor model, which decomposes the factor loading matrix into the component that can be explained by subject-specific covariates and the orthogonal residual component. The covariates' effects on the factor loadings are further modeled by the additive model via sieve approximations. By using the newly proposed Projected-PCA, the rates of convergence of the smooth factor loading matrices are obtained, which are much faster than those of the conventional factor analysis. The convergence is achieved even when the sample size is finite and is particularly appealing in the high-dimension-low-sample-size situation. This leads us to developing nonparametric tests on whether observed covariates have explaining powers on the loadings and whether they fully explain the loadings. The proposed method is illustrated by both simulated data and the returns of the components of the S&P 500 index. |
Persistent Identifier | http://hdl.handle.net/10722/303508 |
ISSN | 2021 Impact Factor: 4.904 2020 SCImago Journal Rankings: 5.877 |
PubMed Central ID | |
ISI Accession Number ID |
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Fan, Jianqing | - |
dc.contributor.author | Liao, Yuan | - |
dc.contributor.author | Wang, Weichen | - |
dc.date.accessioned | 2021-09-15T08:25:27Z | - |
dc.date.available | 2021-09-15T08:25:27Z | - |
dc.date.issued | 2016 | - |
dc.identifier.citation | Annals of Statistics, 2016, v. 44, n. 1, p. 219-254 | - |
dc.identifier.issn | 0090-5364 | - |
dc.identifier.uri | http://hdl.handle.net/10722/303508 | - |
dc.description.abstract | This paper introduces a Projected Principal Component Analysis (Projected-PCA), which employs principal component analysis to the projected (smoothed) data matrix onto a given linear space spanned by covariates. When it applies to high-dimensional factor analysis, the projection removes noise components. We show that the unobserved latent factors can be more accurately estimated than the conventional PCA if the projection is genuine, or more precisely, when the factor loading matrices are related to the projected linear space. When the dimensionality is large, the factors can be estimated accurately even when the sample size is finite. We propose a flexible semiparametric factor model, which decomposes the factor loading matrix into the component that can be explained by subject-specific covariates and the orthogonal residual component. The covariates' effects on the factor loadings are further modeled by the additive model via sieve approximations. By using the newly proposed Projected-PCA, the rates of convergence of the smooth factor loading matrices are obtained, which are much faster than those of the conventional factor analysis. The convergence is achieved even when the sample size is finite and is particularly appealing in the high-dimension-low-sample-size situation. This leads us to developing nonparametric tests on whether observed covariates have explaining powers on the loadings and whether they fully explain the loadings. The proposed method is illustrated by both simulated data and the returns of the components of the S&P 500 index. | - |
dc.language | eng | - |
dc.relation.ispartof | Annals of Statistics | - |
dc.subject | High-dimensionality | - |
dc.subject | Semiparametric factor models | - |
dc.subject | Loading matrix modeling | - |
dc.subject | Sieve approximation | - |
dc.title | Projected principal component analysis in factor models | - |
dc.type | Article | - |
dc.description.nature | link_to_OA_fulltext | - |
dc.identifier.doi | 10.1214/15-AOS1364 | - |
dc.identifier.pmid | 26783374 | - |
dc.identifier.pmcid | PMC4714810 | - |
dc.identifier.scopus | eid_2-s2.0-85013178156 | - |
dc.identifier.volume | 44 | - |
dc.identifier.issue | 1 | - |
dc.identifier.spage | 219 | - |
dc.identifier.epage | 254 | - |
dc.identifier.isi | WOS:000368022000008 | - |