Browse > Article
http://dx.doi.org/10.29220/CSAM.2021.28.1.081

Least quantile squares method for the detection of outliers  

Seo, Han Son (Department of Applied Statistics, Konkuk University)
Yoon, Min (Department of Applied Mathematics, Pukyong National University)
Publication Information
Communications for Statistical Applications and Methods / v.28, no.1, 2021 , pp. 81-88 More about this Journal
Abstract
k-least quantile of squares (k-LQS) estimates are a generalization of least median of squares (LMS) estimates. They have not been used as much as LMS because their breakdown points become small as k increases. But if the size of outliers is assumed to be fixed LQS estimates yield a good fit to the majority of data and residuals calculated from LQS estimates can be a reliable tool to detect outliers. We propose to use LQS estimates for separating a clean set from the data in the context of outlyingness of the cases. Three procedures are suggested for the identification of outliers using LQS estimates. Examples are provided to illustrate the methods. A Monte Carlo study show that proposed methods are effective.
Keywords
influential case; least quantile squares; outliers; masking and swamping effects; regression diagnostics;
Citations & Related Records
연도 인용수 순위
  • Reference
1 Atkinson AC (1994). Fast very robust methods for the detection of multiple outliers, Journal of the American Statistical Association, 89, 1329-1339.   DOI
2 Carrizosa E and Plastria F (1995). The determination of a least quantile of squares regression line for all quantiles, Computational Statistics & Data Analysis, 20, 467-479.   DOI
3 Hadi AS and Simonoff JS (1993). Procedures for the identification of multiple outliers in linear models, Journal of the American Statistical Association, 88, 1264-1272.   DOI
4 Hawkins DM, Bradu D, and Kass GV (1984). Location of several outliers in multiple regression data using elemental sets, Technometrics, 26, 197-208.   DOI
5 Jajo Nethal K (2005). A review of robust regression and diagnostic procedures in linear regression, Acta Mathematicae Applicatae Sinica, 21, 209-224.   DOI
6 Kianifard F and Swallow WH (1989). Using recursive residuals, calculated on adaptive-ordered observations, to identify outliers in linear regression, Biometrics, 45, 571-885.   DOI
7 Kianifard F and Swallow WH (1990). A Monte Carlo comparison of five procedures for identifying outliers in linear regression, Communications in Statistics - Theory and Methods, 19, 1913-1938.   DOI
8 Marasinghe MG (1985). A multistage procedure for detecting several outliers in linear regression, Technometrics, 27, 395-399.   DOI
9 Paul SR and Fung KY (1991). A generalized extreme studentized residual multiple-outlier-detection procedure in linear regression, Technometrics, 33, 339-348.   DOI
10 Pena D and Yohai VJ (1999). A fast procedure for outlier diagnostics in linear regression problems, Journal of the American Statistical Association, 94, 434-445.   DOI
11 Rosner B (1975). On the detection of many outliers, Technometrics, 17, 217-227.
12 Rousseeuw PJ (1984). Least median of squares regression, Journal of the American Statistical Association, 79, 871-880.   DOI
13 Hartigan JA (1981). Consistency of single linkage for high-density clusters, Journal of the American Statistical Association, 76, 388-394.   DOI
14 Rousseeuw PJ and Leroy AM (1987). Robust Regression and Outlier Detection, John Wiley & Sons, New York.
15 Simonoff JS (1988). Detecting outlying cells in two-way contingency tables via backwards-stepping, Technometrics, 30, 339-345.   DOI
16 Stromberg AJ (1993). Computing the exact least median of squares estimate and stability diagnostics in multiple linear regression, SIAM Journal on Scientific Computing, 14, 1289-1299.   DOI
17 Yohai VJ (1987). High breakdown-point and high efficiency robust estimates for regression, Annals of Statistics, 15, 642-656.   DOI
18 Watson GA (1998). On computing the least quantile of squares estimate, SIAM Journal on Scientific Computing, 19, 1125-1138.   DOI