Browse > Article
http://dx.doi.org/10.5351/CSAM.2014.21.5.435

The Bandwidth from the Density Power Divergence  

Pak, Ro Jin (Department of Applied Statistics, Dankook University)
Publication Information
Communications for Statistical Applications and Methods / v.21, no.5, 2014 , pp. 435-444 More about this Journal
Abstract
The most widely used optimal bandwidth is known to minimize the mean integrated squared error(MISE) of a kernel density estimator from a true density. In this article proposes, we propose a bandwidth which asymptotically minimizes the mean integrated density power divergence(MIDPD) between a true density and a corresponding kernel density estimator. An approximated form of the mean integrated density power divergence is derived and a bandwidth is obtained as a product of minimization based on the approximated form. The resulting bandwidth resembles the optimal bandwidth by Parzen (1962), but it reflects the nature of a model density more than the existing optimal bandwidths. We have one more choice of an optimal bandwidth with a firm theoretical background; in addition, an empirical study we show that the bandwidth from the mean integrated density power divergence can produce a density estimator fitting a sample better than the bandwidth from the mean integrated squared error.
Keywords
Density estimator; density power divergence; Kullback-Leibler divergence; $L_2$ distance; mean integrated square error;
Citations & Related Records
연도 인용수 순위
  • Reference
1 Parzen, E. (1962). On estimation of a probability density function and mode, The Annals of Mathematical Statistics, 33, 1065-1076.   DOI   ScienceOn
2 Basu, A., Mandal, A., Martin, N. and Pardo, L. (2013). Testing statistical hypotheses based on the density power divergence, Annals of the Institute of Statistical Mathematics, 65, 319-348.   DOI
3 Devroye, L. and Gyorfi, L. (1985). Nonparametric Density Estimation: The L1 View, Wiley, New York.
4 Durio, A. and Isaia, E. D. (2011). The Minimum density power divergence approach in building robust regression models, Informatica, 22, 43-56.
5 Fujisawa, H. and, Eguchi, F. (2006). Robust estimation in the normal mixture model, Journal of Statistical Planning and Inference, 136, 3989-4011.   DOI
6 Hall, P. (1987). On Kullback-Leibler loss and density estimation, The Annals of Statistics, 15, 1491-1519.   DOI
7 Kanazawa, Y. (1993). Hellinger distance and Kullback-Leiber loss for the kernel density estimator, Statistics and Probability Letters, 18, 315-321.   DOI
8 Kincaid, D. and Cheney, W. (1991). Numerical Analysis: Mathematics of Scientific Computing, Brooks/Cole, New York.
9 Lee, S. and Na, O. (2005). Test for parameter change based on the estimator minimizing density-based divergence measures, Annals of the Institute of Statistical Mathematics, 57, 553-573.   DOI
10 Marron, J. and Wand, M. (1992). Exact mean integrated squared error, The Annals of Statistics, 20, 712-736.   DOI   ScienceOn
11 Rosenblatt, M. (1956). Remarks on some nonparametric estimates of a density function, The Annals of Mathematical Statistics, 27, 832-837.   DOI   ScienceOn
12 Silverman, B.W. (1985). Density Estimation for Statistics and Data Analysis, Chapman and Hall\CRC, New York.
13 Warwick, J. and Jones, M. C. J. (2005). Choosing a robustness tuning parameter, Journal of Statistical Computation and Simulation, 75, 581-588.   DOI
14 Basu, A., Harris, I. R., Hjort, N. L. and Jones, M.C. (1998). Robust and efficient estimation by minimizing a density power divergence, Biometrika, 85, 549-559.   DOI   ScienceOn