• Title/Summary/Keyword: Bayesian statistical method

Search Result 308, Processing Time 0.021 seconds

Model-Based Survival Estimates of Female Breast Cancer Data

  • Khan, Hafiz Mohammad Rafiqullah;Saxena, Anshul;Gabbidon, Kemesha;Rana, Sagar;Ahmed, Nasar Uddin
    • Asian Pacific Journal of Cancer Prevention
    • /
    • v.15 no.6
    • /
    • pp.2893-2900
    • /
    • 2014
  • Background: Statistical methods are very important to precisely measure breast cancer patient survival times for healthcare management. Previous studies considered basic statistics to measure survival times without incorporating statistical modeling strategies. The objective of this study was to develop a data-based statistical probability model from the female breast cancer patients' survival times by using the Bayesian approach to predict future inferences of survival times. Materials and Methods: A random sample of 500 female patients was selected from the Surveillance Epidemiology and End Results cancer registry database. For goodness of fit, the standard model building criteria were used. The Bayesian approach is used to obtain the predictive survival times from the data-based Exponentiated Exponential Model. Markov Chain Monte Carlo method was used to obtain the summary results for predictive inference. Results: The highest number of female breast cancer patients was found in California and the lowest in New Mexico. The majority of them were married. The mean (SD) age at diagnosis (in years) was 60.92 (14.92). The mean (SD) survival time (in months) for female patients was 90.33 (83.10). The Exponentiated Exponential Model found better fits for the female survival times compared to the Exponentiated Weibull Model. The Bayesian method is used to obtain predictive inference for future survival times. Conclusions: The findings with the proposed modeling strategy will assist healthcare researchers and providers to precisely predict future survival estimates as the recent growing challenges of analyzing healthcare data have created new demand for model-based survival estimates. The application of Bayesian will produce precise estimates of future survival times.

Texture Segmentation Using Statistical Characteristics of SOM and Multiscale Bayesian Image Segmentation Technique (SOM의 통계적 특성과 다중 스케일 Bayesian 영상 분할 기법을 이용한 텍스쳐 분할)

  • Kim Tae-Hyung;Eom Il-Kyu;Kim Yoo-Shin
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.42 no.6
    • /
    • pp.43-54
    • /
    • 2005
  • This paper proposes a novel texture segmentation method using Bayesian image segmentation method and SOM(Self Organization feature Map). Multi-scale wavelet coefficients are used as the input of SOM, and likelihood and a posterior probability for observations are obtained from trained SOMs. Texture segmentation is performed by a posterior probability from trained SOMs and MAP(Maximum A Posterior) classification. And the result of texture segmentation is improved by context information. This proposed segmentation method shows better performance than segmentation method by HMT(Hidden Markov Tree) model. The texture segmentation results by SOM and multi-sclae Bayesian image segmentation technique called HMTseg also show better performance than by HMT and HMTseg.

Bayesian Test of Quasi-Independence in a Sparse Two-Way Contingency Table

  • Kwak, Sang-Gyu;Kim, Dal-Ho
    • Communications for Statistical Applications and Methods
    • /
    • v.19 no.3
    • /
    • pp.495-500
    • /
    • 2012
  • We consider a Bayesian test of independence in a two-way contingency table that has some zero cells. To do this, we take a three-stage hierarchical Bayesian model under each hypothesis. For prior, we use Dirichlet density to model the marginal cell and each cell probabilities. Our method does not require complicated computation such as a Metropolis-Hastings algorithm to draw samples from each posterior density of parameters. We draw samples using a Gibbs sampler with a grid method. For complicated posterior formulas, we apply the Monte-Carlo integration and the sampling important resampling algorithm. We compare the values of the Bayes factor with the results of a chi-square test and the likelihood ratio test.

Bayesian Analysis of Multivariate Threshold Animal Models Using Gibbs Sampling

  • Lee, Seung-Chun;Lee, Deukhwan
    • Journal of the Korean Statistical Society
    • /
    • v.31 no.2
    • /
    • pp.177-198
    • /
    • 2002
  • The estimation of variance components or variance ratios in linear model is an important issue in plant or animal breeding fields, and various estimation methods have been devised to estimate variance components or variance ratios. However, many traits of economic importance in those fields are observed as dichotomous or polychotomous outcomes. The usual estimation methods might not be appropriate for these cases. Recently threshold linear model is considered as an important tool to analyze discrete traits specially in animal breeding field. In this note, we consider a hierarchical Bayesian method for the threshold animal model. Gibbs sampler for making full Bayesian inferences about random effects as well as fixed effects is described to analyze jointly discrete traits and continuous traits. Numerical example of the model with two discrete ordered categorical traits, calving ease of calves from born by heifer and calving ease of calf from born by cow, and one normally distributed trait, birth weight, is provided.

Bayesian Semi-Parametric Regression for Quantile Residual Lifetime

  • Park, Taeyoung;Bae, Wonho
    • Communications for Statistical Applications and Methods
    • /
    • v.21 no.4
    • /
    • pp.285-296
    • /
    • 2014
  • The quantile residual life function has been effectively used to interpret results from the analysis of the proportional hazards model for censored survival data; however, the quantile residual life function is not always estimable with currently available semi-parametric regression methods in the presence of heavy censoring. A parametric regression approach may circumvent the difficulty of heavy censoring, but parametric assumptions on a baseline hazard function can cause a potential bias. This article proposes a Bayesian semi-parametric regression approach for inference on an unknown baseline hazard function while adjusting for available covariates. We consider a model-based approach but the proposed method does not suffer from strong parametric assumptions, enjoying a closed-form specification of the parametric regression approach without sacrificing the flexibility of the semi-parametric regression approach. The proposed method is applied to simulated data and heavily censored survival data to estimate various quantile residual lifetimes and adjust for important prognostic factors.

Semiparametric Regression Splines in Matched Case-Control Studies

  • Kim, In-Young;Carroll, Raymond J.;Cohen, Noah
    • Proceedings of the Korean Statistical Society Conference
    • /
    • 2003.05a
    • /
    • pp.167-170
    • /
    • 2003
  • We develop semiparametric methods for matched case-control studies using regression splines. Three methods are developed: an approximate crossvalidation scheme to estimate the smoothing parameter inherent in regression splines, as well as Monte Carlo Expectation Maximization (MCEM) and Bayesian methods to fit the regression spline model. We compare the approximate cross-validation approach, MCEM and Bayesian approaches using simulation, showing that they appear approximately equally efficient, with the approximate cross-validation method being computationally the most convenient. An example from equine epidemiology that motivated the work is used to demonstrate our approaches.

  • PDF

Bayesian inference of the cumulative logistic principal component regression models

  • Kyung, Minjung
    • Communications for Statistical Applications and Methods
    • /
    • v.29 no.2
    • /
    • pp.203-223
    • /
    • 2022
  • We propose a Bayesian approach to cumulative logistic regression model for the ordinal response based on the orthogonal principal components via singular value decomposition considering the multicollinearity among predictors. The advantage of the suggested method is considering dimension reduction and parameter estimation simultaneously. To evaluate the performance of the proposed model we conduct a simulation study with considering a high-dimensional and highly correlated explanatory matrix. Also, we fit the suggested method to a real data concerning sprout- and scab-damaged kernels of wheat and compare it to EM based proportional-odds logistic regression model. Compared to EM based methods, we argue that the proposed model works better for the highly correlated high-dimensional data with providing parameter estimates and provides good predictions.

Complex Segregation Analysis of Categorical Traits in Farm Animals: Comparison of Linear and Threshold Models

  • Kadarmideen, Haja N.;Ilahi, H.
    • Asian-Australasian Journal of Animal Sciences
    • /
    • v.18 no.8
    • /
    • pp.1088-1097
    • /
    • 2005
  • Main objectives of this study were to investigate accuracy, bias and power of linear and threshold model segregation analysis methods for detection of major genes in categorical traits in farm animals. Maximum Likelihood Linear Model (MLLM), Bayesian Linear Model (BALM) and Bayesian Threshold Model (BATM) were applied to simulated data on normal, categorical and binary scales as well as to disease data in pigs. Simulated data on the underlying normally distributed liability (NDL) were used to create categorical and binary data. MLLM method was applied to data on all scales (Normal, categorical and binary) and BATM method was developed and applied only to binary data. The MLLM analyses underestimated parameters for binary as well as categorical traits compared to normal traits; with the bias being very severe for binary traits. The accuracy of major gene and polygene parameter estimates was also very low for binary data compared with those for categorical data; the later gave results similar to normal data. When disease incidence (on binary scale) is close to 50%, segregation analysis has more accuracy and lesser bias, compared to diseases with rare incidences. NDL data were always better than categorical data. Under the MLLM method, the test statistics for categorical and binary data were consistently unusually very high (while the opposite is expected due to loss of information in categorical data), indicating high false discovery rates of major genes if linear models are applied to categorical traits. With Bayesian segregation analysis, 95% highest probability density regions of major gene variances were checked if they included the value of zero (boundary parameter); by nature of this difference between likelihood and Bayesian approaches, the Bayesian methods are likely to be more reliable for categorical data. The BATM segregation analysis of binary data also showed a significant advantage over MLLM in terms of higher accuracy. Based on the results, threshold models are recommended when the trait distributions are discontinuous. Further, segregation analysis could be used in an initial scan of the data for evidence of major genes before embarking on molecular genome mapping.

Optimal Software Release Policy for Random Cost Model

  • Kim, Hee-Soo;Shin, Mi-Young;Park, Dong-Ho
    • Communications for Statistical Applications and Methods
    • /
    • v.12 no.3
    • /
    • pp.673-682
    • /
    • 2005
  • In this paper, we generalize the software reliability growth model by assuming that the testing cost and maintenance cost are random and adopt the Bayesian approach to determine the optimal software release time. Numerical examples are provided to illustrate the Bayesian method for certain parametric models.

Bayes Estimation of a Reliability Function for Rayleigh Model

  • Kim, Yeung-Hoon;Sohn, Joong-Kweon
    • Journal of the Korean Statistical Society
    • /
    • v.23 no.2
    • /
    • pp.445-461
    • /
    • 1994
  • This paper deals with the problem of obtaining some Bayes estimators and Bayesian credible regions of a reliability function for the Rayleigh distribution. Using several priors for a reliability function some Bayes estimators and Bayes credible sets are proposed and studied under squared error loss and Harris loss. Also the performances and behaviors of the proposed Bayes estimators are examined via Monte Carlo simulations and some numericla examples are given.

  • PDF