Browse > Article
http://dx.doi.org/10.29220/CSAM.2022.29.5.615

Intensive comparison of semi-parametric and non-parametric dimension reduction methods in forward regression  

Shin, Minju (Department of Statistics, Ewha Womans University)
Yoo, Jae Keun (Department of Statistics, Ewha Womans University)
Publication Information
Communications for Statistical Applications and Methods / v.29, no.5, 2022 , pp. 615-627 More about this Journal
Abstract
Principal Fitted Component (PFC) is a semi-parametric sufficient dimension reduction (SDR) method, which is originally proposed in Cook (2007). According to Cook (2007), the PFC has a connection with other usual non-parametric SDR methods. The connection is limited to sliced inverse regression (Li, 1991) and ordinary least squares. Since there is no direct comparison between the two approaches in various forward regressions up to date, a practical guidance between the two approaches is necessary for usual statistical practitioners. To fill this practical necessity, in this paper, we newly derive a connection of the PFC to covariance methods (Yin and Cook, 2002), which is one of the most popular SDR methods. Also, intensive numerical studies have done closely to examine and compare the estimation performances of the semi- and non-parametric SDR methods for various forward regressions. The founding from the numerical studies are confirmed in a real data example.
Keywords
covariance methods; principal fitted component; semi-parametric dimension reduction; sufficient dimension reduction;
Citations & Related Records
연도 인용수 순위
  • Reference
1 Cook RD and Weisberg S (1991). Comment: Sliced inverse regression for dimension reduction by KC Li, Journal of the American Statistical Association, 86, 328-332.
2 Fearn T (1983). A misuse of ridge regression in the calibration of a near infrared reflectance instrument, Journal of the Royal Statistical Society: Series C (Applied Statistics), 32, 73-79.   DOI
3 Li L and Yin X (2008) Sliced inverse regression with regularizations, Biometrics, 64, 124-131.   DOI
4 Yoo JK (2018a). Response dimension reduction: model-based approach, Statistics, 52, 409-425.   DOI
5 Cook RD (1998). Regression Graphics, John Wiley & Sons, New York.
6 Cook RD (2007). Fisher lecture: dimension reduction in regression, Statistical Science, 22, 1-26.   DOI
7 Cook RD and Forzani L (2009). Principal fitted components for dimension reduction in regression, Statistical Science, 485, 485-501.
8 Cook RD, Li B, and Chiaromonte F (2007). Dimension reduction in regression without matrix inversion, Biometrika, 94, 569-584.   DOI
9 Cook RD and Weisberg S (1994). An Introduction to Regression Graphics, John Wiley & Sons, New York.
10 Cook RD and Zhang X (2014). Fused estimators of the central subspace in sufficient dimension reduction, Journal of the American Statistical Association, 109, 815-827.   DOI
11 Yoo JK (2016). Tutorial: Methodologies for sufficient dimension reduction in regression, Communications for Statistical Applications and Methods, 23, 95-117.
12 Li KC (1991). Sliced inverse regression for dimension reduction, Journal of the American Statistical Association, 86, 316-327.   DOI
13 Li KC (1992). On principal Hessian directions for data visualization and dimension reduction: another application of Stein's lemma, Journal of the American Statistical Association, 87, 1025-1039.   DOI
14 Yin X and Cook RD (2002). Dimension reduction for the conditional kth moment in regression, Journal of Royal Statistical Society Series B, 64, 159-175.   DOI
15 Yoo JK (2018b). Partial least squares fusing unsupervised learning, Chemometrics and Intelligent Laboratory Systems, 175, 82-86.   DOI