• Title/Summary/Keyword: Poisson count data

Search Result 83, Processing Time 0.03 seconds

Effects of Overdispersion on Testing for Serial Dependence in the Time Series of Counts Data

  • Kim, Hee-Young;Park, You-Sung
    • Communications for Statistical Applications and Methods
    • /
    • v.17 no.6
    • /
    • pp.829-843
    • /
    • 2010
  • To test for the serial dependence in time series of counts data, Jung and Tremayne (2003) evaluated the size and power of several tests under the class of INARMA models based on binomial thinning operations for Poisson marginal distributions. The overdispersion phenomenon(i.e., a variance greater than the expectation) is common in the real world. Overdispersed count data can be modeled by using alternative thinning operations such as random coefficient thinning, iterated thinning, and quasi-binomial thinning. Such thinning operations can lead to time series models of counts with negative binomial or generalized Poisson marginal distributions. This paper examines whether the test statistics used by Jung and Tremayne (2003) on serial dependence in time series of counts data are affected by overdispersion.

Bayesian analysis for the bivariate Poisson regression model: Applications to road safety countermeasures

  • Choe, Hyeong-Gu;Lim, Joon-Beom;Won, Yong-Ho;Lee, Soo-Beom;Kim, Seong-W.
    • Journal of the Korean Data and Information Science Society
    • /
    • v.23 no.4
    • /
    • pp.851-858
    • /
    • 2012
  • We consider a bivariate Poisson regression model to analyze discrete count data when two dependent variables are present. We estimate the regression coefficients as sociated with several safety countermeasures. We use Markov chain and Monte Carlo techniques to execute some computations. A simulation and real data analysis are performed to demonstrate model fitting performances of the proposed model.

Analysis of Marginal Count Failure Data by using Covariates

  • Karim, Md.Rezaul;Suzuki, Kazuyuki
    • International Journal of Reliability and Applications
    • /
    • v.4 no.2
    • /
    • pp.79-95
    • /
    • 2003
  • Manufacturers collect and analyze field reliability data to enhance the quality and reliability of their products and to improve customer satisfaction. To reduce the data collecting and maintenance costs, the amount of data maintained for evaluating product quality and reliability should be minimized. With this in mind, some industrial companies assemble warranty databases by gathering data from different sources for a particular time period. This “marginal count failure data” does not provide (i) the number of failures by when the product entered service, (ii) the number of failures by product age, or (iii) information about the effects of the operating season or environment. This article describes a method for estimating age-based claim rates from marginal count failure data. It uses covariates to identify variations in claims relative to variables such as manufacturing characteristics, time of manufacture, operating season or environment. A Poisson model is presented, and the method is illustrated using warranty claims data for two electrical products.

  • PDF

Sample size calculations for clustered count data based on zero-inflated discrete Weibull regression models

  • Hanna Yoo
    • Communications for Statistical Applications and Methods
    • /
    • v.31 no.1
    • /
    • pp.55-64
    • /
    • 2024
  • In this study, we consider the sample size determination problem for clustered count data with many zeros. In general, zero-inflated Poisson and binomial models are commonly used for zero-inflated data; however, in real data the assumptions that should be satisfied when using each model might be violated. We calculate the required sample size based on a discrete Weibull regression model that can handle both underdispersed and overdispersed data types. We use the Monte Carlo simulation to compute the required sample size. With our proposed method, a unified model with a low failure risk can be used to cope with the dispersed data type and handle data with many zeros, which appear in groups or clusters sharing a common variation source. A simulation study shows that our proposed method provides accurate results, revealing that the sample size is affected by the distribution skewness, covariance structure of covariates, and amount of zeros. We apply our method to the pancreas disorder length of the stay data collected from Western Australia.

Ex-ante and Ex-post Economic Value Analysis on Ecological River Restoration Project (생태하천복원사업 전후 경제적 가치 비교분석)

  • Lee, Yoon;Chang, Hoon;Yoon, Taeyeon;Chung, Young-Keun;Park, Heeyoung
    • Journal of the Korean Regional Science Association
    • /
    • v.31 no.3
    • /
    • pp.39-54
    • /
    • 2015
  • To assess an economic value of Cheonggyecheon river restoration project, an in-depth exit survey data was collected to apply travel cost method in this study. Poisson model, Negative Binomial, Zero-truncated Poisson, and Zero-truncated Negative Binomial model were executed due to the nature of count data. Empirical results showed that regressors were statistically significant and corresponded to general consumer theory. Since our survey data showed over-dispersion, Zero-truncated Negative Binomial was selected as an optimal one to analyze travel demand of Cheonggyecheon by model goodness of fit test among those aforementioned empirical models. Estimating an economic value of Cheonggyecheon river restoration project, which is known as an ecological river restoration project, we used annual visit of individual traveler and an optimal model. Suffice to say that the annual economic value of Cheonggyecheon river restoration project was estimated as 193.4 billion won in 2013.

Variable selection in Poisson HGLMs using h-likelihoood

  • Ha, Il Do;Cho, Geon-Ho
    • Journal of the Korean Data and Information Science Society
    • /
    • v.26 no.6
    • /
    • pp.1513-1521
    • /
    • 2015
  • Selecting relevant variables for a statistical model is very important in regression analysis. Recently, variable selection methods using a penalized likelihood have been widely studied in various regression models. The main advantage of these methods is that they select important variables and estimate the regression coefficients of the covariates, simultaneously. In this paper, we propose a simple procedure based on a penalized h-likelihood (HL) for variable selection in Poisson hierarchical generalized linear models (HGLMs) for correlated count data. For this we consider three penalty functions (LASSO, SCAD and HL), and derive the corresponding variable-selection procedures. The proposed method is illustrated using a practical example.

A generalized regime-switching integer-valued GARCH(1, 1) model and its volatility forecasting

  • Lee, Jiyoung;Hwang, Eunju
    • Communications for Statistical Applications and Methods
    • /
    • v.25 no.1
    • /
    • pp.29-42
    • /
    • 2018
  • We combine the integer-valued GARCH(1, 1) model with a generalized regime-switching model to propose a dynamic count time series model. Our model adopts Markov-chains with time-varying dependent transition probabilities to model dynamic count time series called the generalized regime-switching integer-valued GARCH(1, 1) (GRS-INGARCH(1, 1)) models. We derive a recursive formula of the conditional probability of the regime in the Markov-chain given the past information, in terms of transition probabilities of the Markov-chain and the Poisson parameters of the INGARCH(1, 1) process. In addition, we also study the forecasting of the Poisson parameter as well as the cumulative impulse response function of the model, which is a measure for the persistence of volatility. A Monte-Carlo simulation is conducted to see the performances of volatility forecasting and behaviors of cumulative impulse response coefficients as well as conditional maximum likelihood estimation; consequently, a real data application is given.

Threshold-asymmetric volatility models for integer-valued time series

  • Kim, Deok Ryun;Yoon, Jae Eun;Hwang, Sun Young
    • Communications for Statistical Applications and Methods
    • /
    • v.26 no.3
    • /
    • pp.295-304
    • /
    • 2019
  • This article deals with threshold-asymmetric volatility models for over-dispersed and zero-inflated time series of count data. We introduce various threshold integer-valued autoregressive conditional heteroscedasticity (ARCH) models as incorporating over-dispersion and zero-inflation via conditional Poisson and negative binomial distributions. EM-algorithm is used to estimate parameters. The cholera data from Kolkata in India from 2006 to 2011 is analyzed as a real application. In order to construct the threshold-variable, both local constant mean which is time-varying and grand mean are adopted. It is noted via a data application that threshold model as an asymmetric version is useful in modelling count time series volatility.

Optimal designs for small Poisson regression experiments using second-order asymptotic

  • Mansour, S. Mehr;Niaparast, M.
    • Communications for Statistical Applications and Methods
    • /
    • v.26 no.6
    • /
    • pp.527-538
    • /
    • 2019
  • This paper considers the issue of obtaining the optimal design in Poisson regression model when the sample size is small. Poisson regression model is widely used for the analysis of count data. Asymptotic theory provides the basis for making inference on the parameters in this model. However, for small size experiments, asymptotic approximations, such as unbiasedness, may not be valid. Therefore, first, we employ the second order expansion of the bias of the maximum likelihood estimator (MLE) and derive the mean square error (MSE) of MLE to measure the quality of an estimator. We then define DM-optimality criterion, which is based on a function of the MSE. This criterion is applied to obtain locally optimal designs for small size experiments. The effect of sample size on the obtained designs are shown. We also obtain locally DM-optimal designs for some special cases of the model.

Developing Rear-End Collision Models of Roundabouts in Korea (국내 회전교차로의 추돌사고 모형 개발)

  • Park, Byung Ho;Beak, Tae Hun
    • Journal of the Korean Society of Safety
    • /
    • v.29 no.6
    • /
    • pp.151-157
    • /
    • 2014
  • This study deals with the rear-end collision at roundabouts. The purpose of this study is to develop the accident models of rear-end collision in Korea. In pursuing the above, this study gives particular attention to developing the appropriate models using Poisson, negative binomial model, ZAM, multiple linear and nonlinear regression models, and statistical analysis tools. The main results are as follows. First, the Vuong statistics and overdispersion parameters indicate that ZIP is the most appropriate model among count data models. Second, RMSE, MPB, MAD and correlation coefficient tests show that the multiple nonlinear model is the most suitable to the rear-end collision data. Finally, such the independent variables as traffic volume, ratio of heavy vehicle, number of circulatory roadway lane, number of crosswalk and stop line are adopted in the optimal model.