Browse > Article
http://dx.doi.org/10.5351/CSAM.2016.23.2.093

Tutorial: Dimension reduction in regression with a notion of sufficiency  

Yoo, Jae Keun (Department of Statistics, Ewha Womans University)
Publication Information
Communications for Statistical Applications and Methods / v.23, no.2, 2016 , pp. 93-103 More about this Journal
Abstract
In the paper, we discuss dimension reduction of predictors ${\mathbf{X}}{\in}{{\mathbb{R}}^p}$ in a regression of $Y{\mid}{\mathbf{X}}$ with a notion of sufficiency that is called sufficient dimension reduction. In sufficient dimension reduction, the original predictors ${\mathbf{X}}$ are replaced by its lower-dimensional linear projection without loss of information on selected aspects of the conditional distribution. Depending on the aspects, the central subspace, the central mean subspace and the central $k^{th}$-moment subspace are defined and investigated as primary interests. Then the relationships among the three subspaces and the changes in the three subspaces for non-singular transformation of ${\mathbf{X}}$ are studied. We discuss the two conditions to guarantee the existence of the three subspaces that constrain the marginal distribution of ${\mathbf{X}}$ and the conditional distribution of $Y{\mid}{\mathbf{X}}$. A general approach to estimate them is also introduced along with an explanation for conditions commonly assumed in most sufficient dimension reduction methodologies.
Keywords
central subspace; central $k^{th}$-moment subspace; central mean subspace; dimension reduction subspace; regression; sufficient dimension reduction;
Citations & Related Records
연도 인용수 순위
  • Reference
1 Bickel P (1982). On adaptive estimation, Annals of Statistics, 10, 647-671.   DOI
2 Cook RD (1998). Regression Graphics: Ideas for Studying Regressions through Graphics, Wiley, New York.
3 Cook RD (2000). Save: a method for dimension reduction and graphics in regression, Communications in Statistics-Theory and Methods, 29, 2109-2121.   DOI
4 Cook RD and Li B (2002). Dimension reduction for conditional mean in regression, Annals of Statistics, 30, 455-474.   DOI
5 Cook RD, Li B, and Chiaromonte F (2007). Dimension reduction in regression without matrix inversion, Biometrika, 94, 569-584.   DOI
6 Cook RD and Weisberg S (1991). Comment: Sliced inverse regression for dimension reduction by KC Li, Journal of the American Statistical Association, 86, 328-332.
7 Eaton ML (1986). A characterization of spherical distributions, Journal of Multivariate Analysis, 20, 272-276.   DOI
8 Hall P and Li KC (1993). On almost linearity of low dimensional projections from high dimensional data, Annals of Statistics, 21, 867-889.   DOI
9 Li KC (1991). Sliced inverse regression for dimension reduction, Journal of the American Statistical Association, 86, 316-327.   DOI
10 Li KC (1992). On principal Hessian directions for data visualization and dimension reduction: another application of Stein's lemma, Journal of the American Statistical Association, 87, 1025-1039.   DOI
11 Li L, Cook RD, and Nachtsheim CJ (2004). Cluster-based estimation for sufficient dimension reduction, Computational Statistics & Data Analysis, 47, 175-193.   DOI
12 Rao CR (1965). Linear Statistical Inference and Its Application, Wiley, New York.
13 Shao Y, Cook RD, and Weisberg S (2006). The linearity condition and adaptive estimation in single-index regressions, Retrieved March 1, 2016, from: http://arxiv.org/pdf/1001.4802v1.pdf
14 Yin X and Cook RD (2002). Dimension reduction for the conditional kth moment in regression, Journal of the Royal Statistical Society Series B, 64, 159-175.   DOI