• 제목/요약/키워드: regression statistics

검색결과 5,255건 처리시간 0.024초

More on directional regression

  • Kim, Kyongwon;Yoo, Jae Keun
    • Communications for Statistical Applications and Methods
    • /
    • 제28권5호
    • /
    • pp.553-562
    • /
    • 2021
  • Directional regression (DR; Li and Wang, 2007) is well-known as an exhaustive sufficient dimension reduction method, and performs well in complex regression models to have linear and nonlinear trends. However, the extension of DR is not well-done upto date, so we will extend DR to accommodate multivariate regression and large p-small n regression. We propose three versions of DR for multivariate regression and discuss how DR is applicable for the latter regression case. Numerical studies confirm that DR is robust to the number of clusters and the choice of hierarchical-clustering or pooled DR.

Logistic Model for Normality by Neural Networks

  • Lee, Jea-Young;Rhee, Seong-Won
    • Journal of the Korean Data and Information Science Society
    • /
    • 제14권1호
    • /
    • pp.119-129
    • /
    • 2003
  • We propose a new logistic regression model of normality curves for normal(diseased) and abnormal(nondiseased) classifications by neural networks in data mining. The fitted logistic regression lines are estimated, interpreted and plotted by the neural network technique. A few goodness-of-fit test statistics for normality are discussed and the performances by the fitted logistic regression lines are conducted.

  • PDF

Diagnostics for Regression with Finite-Order Autoregressive Disturbances

  • Lee, Young-Hoon;Jeong, Dong-Bin;Kim, Soon-Kwi
    • Journal of the Korean Statistical Society
    • /
    • 제31권2호
    • /
    • pp.237-250
    • /
    • 2002
  • Motivated by Cook's (1986) assessment of local influence by investigating the curvature of a surface associated with the overall discrepancy measure, this paper extends this idea to the linear regression model with AR(p) disturbances. Diagnostic for the linear regression models with AR(p) disturbances are discussed when simultaneous perturbations of the response vector are allowed. For the derived criterion, numerical studies demonstrate routine application of this work.

Clustering Observations for Detecting Multiple Outliers in Regression Models

  • Seo, Han-Son;Yoon, Min
    • 응용통계연구
    • /
    • 제25권3호
    • /
    • pp.503-512
    • /
    • 2012
  • Detecting outliers in a linear regression model eventually fails when similar observations are classified differently in a sequential process. In such circumstances, identifying clusters and applying certain methods to the clustered data can prevent a failure to detect outliers and is computationally efficient due to the reduction of data. In this paper, we suggest to implement a clustering procedure for this purpose and provide examples that illustrate the suggested procedure applied to the Hadi-Simonoff (1993) method, reverse Hadi-Simonoff method, and Gentleman-Wilk (1975) method.

Diagnostics of partial regression and partial residual plots

  • Lee, Jea-Young;Choi, Suk-Hwa
    • Journal of the Korean Data and Information Science Society
    • /
    • 제11권1호
    • /
    • pp.73-81
    • /
    • 2000
  • The variance inflation factor can be expressed by the square of the ratio of t-statistics associated with slopes of partial regression and partial residual plots. Disagreement of two sides in the interpretation can be occurred, and we analyze it with some illustrations.

  • PDF

Multiple Deletions in Logistic Regression Models

  • Jung, Kang-Mo
    • Communications for Statistical Applications and Methods
    • /
    • 제16권2호
    • /
    • pp.309-315
    • /
    • 2009
  • We extended the results of Roy and Guria (2008) to multiple deletions in logistic regression models. Since single deletions may not exactly detect outliers or influential observations due to swamping effects and masking effects, it needs multiple deletions. We developed conditional deletion diagnostics which are designed to overcome problems of masking effects. We derived the closed forms for several statistics in logistic regression models. They give useful diagnostics on the statistics.

저항적 포아송 회귀와 활용 (Resistant Poisson Regression and Its Application)

  • 허명회;성내경;임용빈
    • 품질경영학회지
    • /
    • 제33권1호
    • /
    • pp.83-87
    • /
    • 2005
  • For the count response we normally consider Poisson regression model. However, the conventional fitting algorithm for Poisson regression model is not reliable at all when the response variable is measured with sizable contamination. In this study, we propose an alternative fitting algorithm that is resistant to outlying values in response and report a case study in semiconductor industry.

On study for change point regression problems using a difference-based regression model

  • Park, Jong Suk;Park, Chun Gun;Lee, Kyeong Eun
    • Communications for Statistical Applications and Methods
    • /
    • 제26권6호
    • /
    • pp.539-556
    • /
    • 2019
  • This paper derive a method to solve change point regression problems via a process for obtaining consequential results using properties of a difference-based intercept estimator first introduced by Park and Kim (Communications in Statistics - Theory Methods, 2019) for outlier detection in multiple linear regression models. We describe the statistical properties of the difference-based regression model in a piecewise simple linear regression model and then propose an efficient algorithm for change point detection. We illustrate the merits of our proposed method in the light of comparison with several existing methods under simulation studies and real data analysis. This methodology is quite valuable, "no matter what regression lines" and "no matter what the number of change points".

Principal Component Regression by Principal Component Selection

  • Lee, Hosung;Park, Yun Mi;Lee, Seokho
    • Communications for Statistical Applications and Methods
    • /
    • 제22권2호
    • /
    • pp.173-180
    • /
    • 2015
  • We propose a selection procedure of principal components in principal component regression. Our method selects principal components using variable selection procedures instead of a small subset of major principal components in principal component regression. Our procedure consists of two steps to improve estimation and prediction. First, we reduce the number of principal components using the conventional principal component regression to yield the set of candidate principal components and then select principal components among the candidate set using sparse regression techniques. The performance of our proposals is demonstrated numerically and compared with the typical dimension reduction approaches (including principal component regression and partial least square regression) using synthetic and real datasets.

Iterative projection of sliced inverse regression with fused approach

  • Han, Hyoseon;Cho, Youyoung;Yoo, Jae Keun
    • Communications for Statistical Applications and Methods
    • /
    • 제28권2호
    • /
    • pp.205-215
    • /
    • 2021
  • Sufficient dimension reduction is useful dimension reduction tool in regression, and sliced inverse regression (Li, 1991) is one of the most popular sufficient dimension reduction methodologies. In spite of its popularity, it is known to be sensitive to the number of slices. To overcome this shortcoming, the so-called fused sliced inverse regression is proposed by Cook and Zhang (2014). Unfortunately, the two existing methods do not have the direction application to large p-small n regression, in which the dimension reduction is desperately needed. In this paper, we newly propose seeded sliced inverse regression and seeded fused sliced inverse regression to overcome this deficit by adopting iterative projection approach (Cook et al., 2007). Numerical studies are presented to study their asymptotic estimation behaviors, and real data analysis confirms their practical usefulness in high-dimensional data analysis.