Sparse functional principal component analysis in a new regression framework☆
Introduction
Functional principal component analysis (FPCA) is a crucial dimension reduction tool in functional data analysis. FPCA explores major sources of variability in a sample of random curves by finding functional principal components (FPCs) that maximize the curve variation. Consequently, the top few FPCs explain most of the variability in the random curves. Besides, each random curve can be approximated by a linear combination of the top few FPCs. Therefore, the infinite-dimensional curves are projected to a low-dimensional space defined by the top FPCs. This powerful dimensional reduction feature also promotes the popularity of FPCA.
The theoretical properties of FPCA have been carefully studied at length. For example, Dauxois et al. (1982) first studied the asymptotic properties of PCA estimators for the infinite-dimensional data from a linear operator viewpoint. Following this point of view, Mas (2002) and Bosq (2000) utilized functional analysis to study FPCA theoretically. On the other hand, Hall and Horowitz, 2007, Hall et al., 2006 and Yao et al. (2005a) studied FPCA from the kernel perspective. Sang et al. (2017) proposed a parametric approach for estimating FPCs to enhance their interpretability for users. Nie et al. (2018) propose a supervised version of FPCA by considering the correlation of the functional predictor and response variable. In addition, FPCA has been widely and successfully applied in many applications such as functional linear regression (Yao et al., 2005b), classification and clustering of functional data (Ramsay and Silverman, 2005, Yao et al., 2005b, Müller, 2005, Müller and Stadtmüller, 2005, Peng and Müller, 2008, Dong et al., 2018). All these applications assume the functional data are densely and regularly observed. When it comes to sparse and irregularly observed data, (Yao et al., 2005a) proposed to estimate the FPC score using conditional expectation, which allows recovering the individual trajectory by borrowing information across all the subjects. The smooth version of functional principal component analysis is carefully studied by Rice and Silverman, 1991, Pezzulli, 1993, Silverman, 1996, and Yao et al. (2005a). There are mainly three methods to achieve smoothness. The first method smooths the functional data in the first step and conducts the regular FPCA on the sample covariance function. The second method smooths the covariance function first and then eigendecomposes the resulting smoothed covariance function to estimate the smoothed FPCs. The last method directly adds a roughness penalty in the optimization criterion for estimating the FPCs.
The conventional FPCA aims to estimate FPCs that maximize the curve variation. These FPCs represent the source or direction of maximum variations among curves, and the curves are projected to the low-dimensional space defined by these FPCs. Therefore, it is essential to interpret them. However, these FPCs are usually nonzero in the whole observed domain, and users often find it hard to interpret these FPCs. On the other hand, if the estimated FPC is only nonzero in a subregion of the entire domain, we can easily use them to identify the subregions from which the major variation of the curves exhibits. In this paper, our goal is to propose a method to estimate the sparse functional principal components, which are only nonzero in a subregion and, at the same time, account for an almost maximum amount of variation within the curves.
Several methods have been proposed to enhance the interpretability of functional principal components. The first method is the interpretable functional principal components analysis (iFPCA) proposed by Lin et al. (2016). This method adds an -penalty on the length of the nonzero region of FPCs and obtains FPCs, which are only nonzero in subregions. However, the optimization in their framework is an NP-hard problem because of the use of the -penalty. A greedy backward elimination algorithm is proposed to solve this optimization problem approximately. The second method is called a localized functional principal components analysis (LFPCA) method proposed by Chen and Lei (2015). This method adds an penalty to the original eigendecomposition problem of smoothed FPCs, which is also not a convex optimization problem. They approximate this non-convex problem through a Deflated Fantope Localization method and propose a novel estimation procedure in a sequential manner. In addition, Di et al. (2014) considered the functional principal component analysis on sparsely sampled multilevel functional data. The sparsity in their work refers to the situations when the functional data are not fully observed rather than the sparsity of the FPCs. Li et al. (2016) studied the problem when the low-rank structure of the functional data was related to multivariate supervision data. The resulting supervised FPCs incorporate the information carried within the response data. In comparison, our work needs no supervision information and assumes the underlying FPCs are sparse on their own.
This paper has three major contributions. Firstly, we propose a new regression-type framework for the sparse functional principal component analysis. The estimated sparse FPCs can not only account for a reasonable variation within the functional data but also be sparse on the whole domain. We also show that the FPCs estimated with our proposed sparse FPCA method is equivalent to the FPCs with the conventional FPCA method when the sparsity parameter is zero. Secondly, our approach is not an NP-hard optimization problem, and the computation is very efficient. Lastly, our method estimates the top sparse FPCs simultaneously rather than sequentially estimating each FPC. Sequentially estimating the FPCs often leads to a quadratic optimization problem with multiple linear constraints. The numerical complexity increases as the rank of FPCs increases. Besides, the sequential manner does not allow parallel computing because the th FPC can only be estimated after obtaining the first FPCs. In our regression framework, the regression step of our algorithm only involves individual FPC such that it can be solved in a parallel way. An R package “sparseFPCA” is developed to implement our proposed sparse FPC (SFPCA) method. The computing scripts for our simulation study can be downloaded at https://github.com/caojiguo/sparseFPCA.
The rest of the paper is organized as follows. In Section 2, we introduce our SFPCA method and show its connection with the conventional FPCA. Details of our approach and the computation algorithm are described in Section 3. In Section 4, we apply our proposed method in a real-data application to explore major sources of variation among the acceleration rates of 107 diesel trucks. In Section 5, two carefully-designed simulations are conducted to evaluate the finite sample performance of our proposed method in comparison with other alternative methods in different settings. Section 6 provides concluding remarks.
Section snippets
Sparse functional principal component analysis
Consider a stochastic process , which is square-integrable on the compact domain . In other words, almost surely, where is the Hilbert space of square-integrable functions on . We denote the inner product between two functions as with the corresponding norm as . If , we also use the notation . Let , be the observed functional data for the stochastic process . The rest of the paper assumes that the functional data are
Estimation method
We propose to estimate the first unnormalized sparse FPCs, , in an iterative optimization method. More specifically, within each iteration, the first step is to find the optimal that minimizes the criterion (1) given the current estimate of and the second step is to search for a new which further minimizes the proposed criterion conditional on the optimal from the first step. This procedure is repeated until it converges. In the rest of this section, we first give the
Application
Our proposed method is demonstrated by analyzing a real dataset relating to particulate matter (PM) emissions from diesel trucks (Clark et al., 2007). In the experiment, trucks are driven through a pre-determined driving cycle and PM at the exhaust pipe is measured every second via a particulate matter counter. Hall and Hooker (2016) analyzed this dataset to predict PM using the acceleration rate with a functional linear model. Fig. 1 displays the acceleration rate curves for 107 diesel trucks.
Simulation study
We conduct two simulation studies to evaluate our proposed sparse FPCA method by comparing it with three available methods, including the conventional FPCA method (Ramsay and Silverman, 2005), the interpretable functional principal component analysis (iFPCA) proposed by Lin et al. (2016) and the localized functional principal component analysis (LFPCA) proposed by Chen and Lei (2015).
More specifically, the true underlying functional curves are generated using
Conclusions
In this paper, we focused on the problem of estimating FPCs with compact support. The conventional FPCA method estimate FPCs by maximizing the variation among the functional data. But these estimated FPCs are nonzero in the entire domain. Hence they are often hard to interpret.
Conventional FPCA methods estimate FPCs by eigendecomposing the sample covariance function. However, when we need to add the regulation penalty to the FPCs, this eigendecomposition method always leads to an NP-hard
Acknowledgments
The authors are very grateful for the constructive comments from the Editor, an Associate Editor, and two reviewers. These comments are very helpful for the authors to improve their work. This research was supported by the Postgraduate Scholarship-Doctorial (PGS-D) to Y. Nie from the Natural Sciences and Engineering Research Council of Canada (NSERC) and the NSERC, Canada Discovery grant (RGPIN-2018-06008) of J. Cao.
References (27)
- et al.
Asymptotic theory for the principal component analysis of a vector random function: some applications to statistical inference
J. Multivariate Anal.
(1982) - et al.
Representation of random processes using the finite Karhunen-Loeve expansion
Inf. Control
(1970) Weak convergence for the covariance operators of a Hilbertian linear process
Stochastic Process. Appl.
(2002)Linear Processes in Function Spaces: Theory and Applications
(2000)- et al.
Localized functional principal component analysis
J. Amer. Statist. Assoc.
(2015) - et al.
Heavy-duty vehicle chassis dynamometer testing for emissions inventory, air quality modeling, source apportionment and air toxics emissions inventory
Coordinat. Res. Council
(2007) - et al.
Multilevel sparse functional principal component analysis
Stat
(2014) - et al.
Functional principal component analysis of GFR curves after kidney transplant
Stat. Methods Med. Res.
(2018) - et al.
Variable selection via nonconcave penalized likelihood and its oracle properties
J. Amer. Statist. Assoc.
(2001) - et al.
Truncated linear models for functional data
J. R. Stat. Soc. Ser. B Stat. Methodol.
(2016)
Methodology and convergence rates for functional linear regression
Ann. Statist.
Properties of principal component methods for functional and longitudinal data analysis
Ann. Statist.
Supervised sparse and functional principal component analysis
J. Comput. Graph. Statist.
Cited by (0)
- ☆
An R package “sparseFPCA” is developed to implement the proposed method. The computing scripts for the simulation study can be downloaded at https://github.com/caojiguo/sparseFPCA.