• 제목/요약/키워드: Linear Regression Fit

검색결과 138건 처리시간 0.025초

단순회귀모형에서 선형성 검정통계량 (A linearity test statistic in a simple linear regression)

  • 박천건;이경은
    • Journal of the Korean Data and Information Science Society
    • /
    • 제25권2호
    • /
    • pp.305-315
    • /
    • 2014
  • 전통적으로 단순선형회귀모형에서 설명변수와 반응변수의 선형성 평가는 산점도로 쉽게 파악되었다. 보통 반복수가 존재하는 자료에서 적합결여검정은 선형성을 평가하는데 사용되었다. 하지만 반복수가 오직 하나인 경우에 선형성 검정이 수월하지 않다. 본 연구에서는 반복수가 오직 하나인 단순선형회귀모형의 선형성을 검정하는 통계량을 제안하고 모의실험 및 실증연구를 통하여 신뢰성을 파악한다.

Combination of Schwarz Information Criteria for Change-Point Analysis

  • 김종태
    • Journal of the Korean Data and Information Science Society
    • /
    • 제13권2호
    • /
    • pp.185-193
    • /
    • 2002
  • The purpose of this paper is to suggest a method for detecting the linear regression change-points or variance change-points in regression model by the combination of Schwarz information criteria. The advantage of the suggested method is to detect change-points more detailed when one compares the suggest method with Chen (1998)'s method.

  • PDF

The horizontal line detection method using Haar-like features and linear regression in infrared images

  • Park, Byoung Sun;Kim, Jae Hyup
    • 한국컴퓨터정보학회논문지
    • /
    • 제20권12호
    • /
    • pp.29-36
    • /
    • 2015
  • In this paper, we propose the horizontal line detection using the Haar-like features and linear regression in infrared images. In the marine environment horizon image is very useful information on a variety of systems. In the proposed method Haar-like features it was noted that the standard deviation be calculated in real time on a static area. Based on the pixel position, calculating the standard deviation of the around area in real time and, if the reaction is to filter out the largest pixel can get the energy map of the area containing the straight horizontal line. In order to select a horizontal line of pixels from the energy map, we applied the linear regression, calculating a linear fit to the transverse horizontal line across the image to select the candidate optimal horizontal. The proposed method was carried out in a horizontal line detecting real infrared image experiment for day and night, it was confirmed the excellent detection results than the legacy methods.

On Information Criteria in Linear Regression Model

  • Park, Man-Sik
    • 응용통계연구
    • /
    • 제22권1호
    • /
    • pp.197-204
    • /
    • 2009
  • In the model selection problem, the main objective is to choose the true model from a manageable set of candidate models. An information criterion gauges the validity of a statistical model and judges the balance between goodness-of-fit and parsimony; "how well observed values ran approximate to the true values" and "how much information can be explained by the lower dimensional model" In this study, we introduce some information criteria modified from the Akaike Information Criterion (AIC) and the Bayesian Information Criterion(BIC). The information criteria considered in this study are compared via simulation studies and real application.

An Alternative Method of Regression: Robust Modified Anti-Hebbian Learning

  • Hwang, Chang-Ha
    • Journal of the Korean Data and Information Science Society
    • /
    • 제7권2호
    • /
    • pp.203-210
    • /
    • 1996
  • A linear neural unit with a modified anti-Hebbian learning rule has been shown to be able to optimally fit curves, surfaces, and hypersurfaces by adaptively extracting the minor component of the input data set. In this paper, we study how to use the robust version of this neural fitting method for linear regression analysis. Furthermore, we compare this method with other methods when data set is contaminated by outliers.

  • PDF

Case influence diagnostics for the significance of the linear regression model

  • Bae, Whasoo;Noh, Soyoung;Kim, Choongrak
    • Communications for Statistical Applications and Methods
    • /
    • 제24권2호
    • /
    • pp.155-162
    • /
    • 2017
  • In this paper we propose influence measures for two basic goodness-of-fit statistics, the coefficient of determination $R^2$ and test statistic F in the linear regression model using the deletion method. Some useful lemmas are provided. We also express the influence measures in terms of basic building blocks such as residual, leverage, and deviation that showed them as increasing function of residuals and a decreasing function of deviation. Further, the proposed measure reduces computational burden from O(n) to O(1). As illustrative examples, we applied the proposed measures to the stackloss data sets. We verified that deletion of one or few influential observations may result in big change in $R^2$ and F-statistic.

Wage Determinants Analysis by Quantile Regression Tree

  • Chang, Young-Jae
    • Communications for Statistical Applications and Methods
    • /
    • 제19권2호
    • /
    • pp.293-301
    • /
    • 2012
  • Quantile regression proposed by Koenker and Bassett (1978) is a statistical technique that estimates conditional quantiles. The advantage of using quantile regression is the robustness in response to large outliers compared to ordinary least squares(OLS) regression. A regression tree approach has been applied to OLS problems to fit flexible models. Loh (2002) proposed the GUIDE algorithm that has a negligible selection bias and relatively low computational cost. Quantile regression can be regarded as an analogue of OLS, therefore it can also be applied to GUIDE regression tree method. Chaudhuri and Loh (2002) proposed a nonparametric quantile regression method that blends key features of piecewise polynomial quantile regression and tree-structured regression based on adaptive recursive partitioning. Lee and Lee (2006) investigated wage determinants in the Korean labor market using the Korean Labor and Income Panel Study(KLIPS). Following Lee and Lee, we fit three kinds of quantile regression tree models to KLIPS data with respect to the quantiles, 0.05, 0.2, 0.5, 0.8, and 0.95. Among the three models, multiple linear piecewise quantile regression model forms the shortest tree structure, while the piecewise constant quantile regression model has a deeper tree structure with more terminal nodes in general. Age, gender, marriage status, and education seem to be the determinants of the wage level throughout the quantiles; in addition, education experience appears as the important determinant of the wage level in the highly paid group.

Residuals Plots for Repeated Measures Data

  • 박태성
    • 한국통계학회:학술대회논문집
    • /
    • 한국통계학회 2000년도 추계학술발표회 논문집
    • /
    • pp.187-191
    • /
    • 2000
  • In the analysis of repeated measurements, multivariate regression models that account for the correlations among the observations from the same subject are widely used. Like the usual univariate regression models, these multivariate regression models also need some model diagnostic procedures. In this paper, we propose a simple graphical method to detect outliers and to investigate the goodness of model fit in repeated measures data. The graphical method is based on the quantile-quantile(Q-Q) plots of the $X^2$ distribution and the standard normal distribution. We also propose diagnostic measures to detect influential observations. The proposed method is illustrated using two examples.

  • PDF

Effects on Regression Estimates under Misspecified Generalized Linear Mixed Models for Counts Data

  • Jeong, Kwang Mo
    • 응용통계연구
    • /
    • 제25권6호
    • /
    • pp.1037-1047
    • /
    • 2012
  • The generalized linear mixed model(GLMM) is widely used in fitting categorical responses of clustered data. In the numerical approximation of likelihood function the normality is assumed for the random effects distribution; subsequently, the commercial statistical packages also routinely fit GLMM under this normality assumption. We may also encounter departures from the distributional assumption on the response variable. It would be interesting to investigate the impact on the estimates of parameters under misspecification of distributions; however, there has been limited researche on these topics. We study the sensitivity or robustness of the maximum likelihood estimators(MLEs) of GLMM for counts data when the true underlying distribution is normal, gamma, exponential, and a mixture of two normal distributions. We also consider the effects on the MLEs when we fit Poisson-normal GLMM whereas the outcomes are generated from the negative binomial distribution with overdispersion. Through a small scale Monte Carlo study we check the empirical coverage probabilities of parameters and biases of MLEs of GLMM.