Simultaneous dimension reduction and variable selection in modeling high dimensional data

Joseph Ryan G. Lansangan, Erniel B. Barrios

Research output: Contribution to journalArticleResearchpeer-review

12 Citations (Scopus)

Abstract

High dimensional predictors in regression analysis are often associated with multicollinearity along with other estimation problems. These problems can be mitigated through a constrained optimization method that simultaneously induces dimension reduction and variable selection that also maintains a high level of predictive ability of the fitted model. Simulation studies show that the method may outperform sparse principal component regression, least absolute shrinkage and selection operator, and elastic net procedures in terms of predictive ability and optimal selection of inputs. Furthermore, the method yields reduced models with smaller prediction errors than the estimated full models from the principal component regression or the principal covariance regression.

Original languageEnglish
Pages (from-to)242-256
Number of pages15
JournalComputational Statistics and Data Analysis
Volume112
DOIs
Publication statusPublished - Aug 2017
Externally publishedYes

Keywords

  • Dimension reduction
  • High dimensionality
  • Latent factors
  • Regression modeling
  • Soft thresholding
  • Sparse principal component analysis
  • Sparsity
  • Variable selection

Cite this