Dynamic Principal Subspaces with Sparsity in High Dimensions

04/07/2021
by   Xiaoyu Hu, et al.
0

Principal component analysis (PCA) is a versatile tool to reduce the dimensionality which has wide applications in statistics and machine learning community. It is particularly useful to model data in high-dimensional scenarios where the number of variables p is comparable to, or much larger than the sample size n. Despite extensive literature on this topic, researches have focused on modeling static principal eigenvectors or subspaces, which is unsuitable for stochastic processes that are dynamic in nature. To characterize the change in the whole course of high-dimensional data collection, we propose a unified framework to estimate dynamic principal subspaces spanned by leading eigenvectors of covariance matrices. In the proposed framework, we formulate an optimization problem by combining the kernel smoothing and regularization penalty together with the orthogonality constraint, which can be effectively solved by the proximal gradient method for manifold optimization. We show that our method is suitable for high-dimensional data observed under both common and irregular designs. In addition, theoretical properties of the estimators are investigated under l_q (0 ≤ q ≤ 1) sparsity. Extensive experiments demonstrate the effectiveness of the proposed method in both simulated and real data examples.

READ FULL TEXT
research
11/02/2020

Sparse Functional Principal Component Analysis in High Dimensions

Functional principal component analysis (FPCA) is a fundamental tool and...
research
03/27/2019

An Alternating Manifold Proximal Gradient Method for Sparse PCA and Sparse CCA

Sparse principal component analysis (PCA) and sparse canonical correlati...
research
01/22/2019

A Fast Iterative Algorithm for High-dimensional Differential Network

Differential network is an important tool to capture the changes of cond...
research
05/11/2020

Robust PCA via Regularized REAPER with a Matrix-Free Proximal Algorithm

Principal component analysis (PCA) is known to be sensitive to outliers,...
research
02/24/2021

PCA Rerandomization

Mahalanobis distance between treatment group and control group covariate...
research
05/17/2023

Flagfolds

By interpreting the product of the Principal Component Analysis, that is...
research
03/13/2020

A Wide Dataset of Ear Shapes and Pinna-Related Transfer Functions Generated by Random Ear Drawings

Head-related transfer functions (HRTFs) individualization is a key matte...

Please sign up or login with your details

Forgot password? Click here to reset