• Title/Summary/Keyword: Censored Data

Search Result 404, Processing Time 0.026 seconds

Expected Probability Weighted Moment Estimator for Censored Flood Data (절단된 홍수 자료에 대한 확률가중적률 추정량)

  • Jeon, Jong-June;Kim, Young-Oh;Kim, Yong-Dai;Park, June-Hyeong
    • Proceedings of the Korea Water Resources Association Conference
    • /
    • 2010.05a
    • /
    • pp.357-361
    • /
    • 2010
  • 미래의 연별 최대 강수량 예측의 정확성을 향상시키는데 역사적 자료가 도움이 된다는 많은 연구 결과가 있었다. 관측의 오차와 자료의 손실로 역사자료를 이용한 강수 예측 방법은 절단자료의 분석을 중심으로 연구되었다. 대표적인 역사자료의 이용방법으로 조건부 적률을 이용한 B17B [Interagency Committee in Water Data, 1982], 조건부적률과적률 관계식을 이용한 Expected Moment Algorithm(EMA) [Cohn et al.;1997], 조건부 확률가중적률을 이용한 Partial Probability Weighted Moment (PPWM)[Wang ; 1991] 방법이 있다. 본 연구에서는 역사적 자료를 반영하는 방법에 있어 B17B와 EMA의 관계를 밝히고 그러한 관계가 PPWM에 동일하게 적용할 수 있음을 보였다. 우리는 B17B와 EMA의 관계를 적률방정식으로 표현하였고 PPWM에서 확률가중 적률 방정식을 정의함으로써 PPWM을 확장하였다. 본 연구에서 제안한 새로운 역사 자료를 이용한 강수예측 방법론을 Expected Probability Weighted Momemt (EPWM) 방법이라고 부르고 그 예측 방법의 성능을 다른 예측방법과 시뮬레이션 결과를 통해 비교하였다. 역사 자료 방법론의 비교는 Generalized Extreme Value (GEV) 분포를 이용하여 이루어졌으며, 각 방법론은 GEV분포의 형태모수(shape parameter)따라 다른 특성을 나타난다는 것을 보였다. 뿐만 아니라 여기서 제안한 EPWM 방법은 대부분의 경우에 좋은 추정량을 준다는 것을 보였다.

  • PDF

A Case Study of Enhancing Flame Retardancy of Mixture Material (혼합물 소재의 난연성 향상을 위한 실험연구 사례)

  • Byun, Jai-Hyun;Jung, Chun-Sik;Kim, Donghak;Park, Kyu-Hwan
    • Journal of Korean Society for Quality Management
    • /
    • v.47 no.3
    • /
    • pp.631-639
    • /
    • 2019
  • Purpose: In this paper we present a case study of applying quick and easy experimental design approach to develop a Halogen free flame retardant material for cellular phone charger cable. Methods: We employ sequential experimentation of mixture design, verification design, and factorial design. A quick and easy approach is adopted based on data investigation and graphical method instead of strict statistical analysis, which helped enhancing smooth communication with the engineers and speeding up the development process. Results: Flame retardant material in pellet type produced from the optimal condition is transported to the customer and tested, to pass the customer retardancy criteria. Conclusion: The quick and easy experimental design approach is considered to be useful in this case study.

Bayesian and maximum likelihood estimations from exponentiated log-logistic distribution based on progressive type-II censoring under balanced loss functions

  • Chung, Younshik;Oh, Yeongju
    • Communications for Statistical Applications and Methods
    • /
    • v.28 no.5
    • /
    • pp.425-445
    • /
    • 2021
  • A generalization of the log-logistic (LL) distribution called exponentiated log-logistic (ELL) distribution on lines of exponentiated Weibull distribution is considered. In this paper, based on progressive type-II censored samples, we have derived the maximum likelihood estimators and Bayes estimators for three parameters, the survival function and hazard function of the ELL distribution. Then, under the balanced squared error loss (BSEL) and the balanced linex loss (BLEL) functions, their corresponding Bayes estimators are obtained using Lindley's approximation (see Jung and Chung, 2018; Lindley, 1980), Tierney-Kadane approximation (see Tierney and Kadane, 1986) and Markov Chain Monte Carlo methods (see Hastings, 1970; Gelfand and Smith, 1990). Here, to check the convergence of MCMC chains, the Gelman and Rubin diagnostic (see Gelman and Rubin, 1992; Brooks and Gelman, 1997) was used. On the basis of their risks, the performances of their Bayes estimators are compared with maximum likelihood estimators in the simulation studies. In this paper, research supports the conclusion that ELL distribution is an efficient distribution to modeling data in the analysis of survival data. On top of that, Bayes estimators under various loss functions are useful for many estimation problems.

Bayesian bi-level variable selection for genome-wide survival study

  • Eunjee Lee;Joseph G. Ibrahim;Hongtu Zhu
    • Genomics & Informatics
    • /
    • v.21 no.3
    • /
    • pp.28.1-28.13
    • /
    • 2023
  • Mild cognitive impairment (MCI) is a clinical syndrome characterized by the onset and evolution of cognitive impairments, often considered a transitional stage to Alzheimer's disease (AD). The genetic traits of MCI patients who experience a rapid progression to AD can enhance early diagnosis capabilities and facilitate drug discovery for AD. While a genome-wide association study (GWAS) is a standard tool for identifying single nucleotide polymorphisms (SNPs) related to a disease, it fails to detect SNPs with small effect sizes due to stringent control for multiple testing. Additionally, the method does not consider the group structures of SNPs, such as genes or linkage disequilibrium blocks, which can provide valuable insights into the genetic architecture. To address the limitations, we propose a Bayesian bi-level variable selection method that detects SNPs associated with time of conversion from MCI to AD. Our approach integrates group inclusion indicators into an accelerated failure time model to identify important SNP groups. Additionally, we employ data augmentation techniques to impute censored time values using a predictive posterior. We adapt Dirichlet-Laplace shrinkage priors to incorporate the group structure for SNP-level variable selection. In the simulation study, our method outperformed other competing methods regarding variable selection. The analysis of Alzheimer's Disease Neuroimaging Initiative (ADNI) data revealed several genes directly or indirectly related to AD, whereas a classical GWAS did not identify any significant SNPs.

Comparison of parametric and nonparametric hazard change-point estimators (모수적과 비모수적 위험률 변화점 통계량 비교)

  • Kim, Jaehee;Lee, Sieun
    • Journal of the Korean Data and Information Science Society
    • /
    • v.27 no.5
    • /
    • pp.1253-1262
    • /
    • 2016
  • When there exists a change-point in hazard function, it should be estimated for exact parameter or hazard estimation. In this research, we compare the hazard change-point estimators. Matthews and Farewell (1982) parametric change-point estimator is based on the likelihood and Zhang et al. (2014) nonparametric estimator is based on the Nelson-Aalen cumulative hazard estimator. Simulation study is done for the data from exponential distribution with one hazard change-point. The simulated data generated without censoring and the data with right censoring are considered. As real data applications, the change-point estimates are computed for leukemia data and primary biliary cirrhosis data.

Survival Analysis of Gastric Cancer Patients with Incomplete Data

  • Moghimbeigi, Abbas;Tapak, Lily;Roshanaei, Ghodaratolla;Mahjub, Hossein
    • Journal of Gastric Cancer
    • /
    • v.14 no.4
    • /
    • pp.259-265
    • /
    • 2014
  • Purpose: Survival analysis of gastric cancer patients requires knowledge about factors that affect survival time. This paper attempted to analyze the survival of patients with incomplete registered data by using imputation methods. Materials and Methods: Three missing data imputation methods, including regression, expectation maximization algorithm, and multiple imputation (MI) using Monte Carlo Markov Chain methods, were applied to the data of cancer patients referred to the cancer institute at Imam Khomeini Hospital in Tehran in 2003 to 2008. The data included demographic variables, survival times, and censored variable of 471 patients with gastric cancer. After using imputation methods to account for missing covariate data, the data were analyzed using a Cox regression model and the results were compared. Results: The mean patient survival time after diagnosis was $49.1{\pm}4.4$ months. In the complete case analysis, which used information from 100 of the 471 patients, very wide and uninformative confidence intervals were obtained for the chemotherapy and surgery hazard ratios (HRs). However, after imputation, the maximum confidence interval widths for the chemotherapy and surgery HRs were 8.470 and 0.806, respectively. The minimum width corresponded with MI. Furthermore, the minimum Bayesian and Akaike information criteria values correlated with MI (-821.236 and -827.866, respectively). Conclusions: Missing value imputation increased the estimate precision and accuracy. In addition, MI yielded better results when compared with the expectation maximization algorithm and regression simple imputation methods.

Estimation of conditional mean residual life function with random censored data (임의중단자료에서의 조건부 평균잔여수명함수 추정)

  • Lee, Won-Kee;Song, Myung-Unn;Jeong, Seong-Hwa
    • Journal of the Korean Data and Information Science Society
    • /
    • v.22 no.1
    • /
    • pp.89-97
    • /
    • 2011
  • The aims of this study were to propose a method of estimation for mean residual life function (MRLF) from conditional survival function using the Buckley and James's (1979) pseudo random variables, and then to assess the performance of the proposed method through the simulation studies. The mean squared error (MSE) of proposed method were less than those of the Cox's proportional hazard model (PHM) and Beran's nonparametric method for non-PHM case. Futhermore in the case of PHM, the MSE's of proposed method were similar to those of Cox's PHM. Finally, to evaluate the appropriateness of practical use, we applied the proposed method to the gastric cancer data. The data set consist of the 1, 192 patients with gastric cancer underwent surgery at the Department of Surgery, K-University Hospital.

Developing statistical models and constructing clinical systems for analyzing semi-competing risks data produced from medicine, public heath, and epidemiology (의료, 보건, 역학 분야에서 생산되는 준경쟁적 위험자료를 분석하기 위한 통계적 모형의 개발과 임상분석시스템 구축을 위한 연구)

  • Kim, Jinheum
    • The Korean Journal of Applied Statistics
    • /
    • v.33 no.4
    • /
    • pp.379-393
    • /
    • 2020
  • A terminal event such as death may censor an intermediate event such as relapse, but not vice versa in semi-competing risks data, which is often seen in medicine, public health, and epidemiology. We propose a Weibull regression model with a normal frailty to analyze semi-competing risks data when all three transition times of the illness-death model are possibly interval-censored. We construct the conditional likelihood separately depending on the types of subjects: still alive with or without the intermediate event, dead with or without the intermediate event, and dead with the intermediate event missing. Optimal parameter estimates are obtained from the iterative quasi-Newton algorithm after the marginalization of the full likelihood using the adaptive importance sampling. We illustrate the proposed method with extensive simulation studies and PAQUID (Personnes Agées Quid) data.

The wage determinants of college graduates using Heckman's sample selection model (Heckman의 표본선택모형을 이용한 대졸자의 임금결정요인 분석)

  • Cho, Jangsik
    • Journal of the Korean Data and Information Science Society
    • /
    • v.28 no.5
    • /
    • pp.1099-1107
    • /
    • 2017
  • In this study, we analyzed the determinants of wages of college graduates by using the data of "2014 Graduates Occupational Mobility Survey" conducted by Korea Employment Information Service. In general, wages contain two complex pieces of information about whether an individual is employed and the size of the wage. However, in many previous researches on wage determinants, sample selection bias tends to be generated by performing linear regression analysis using only information on wage size. We used the Heckman sample selection models for analysis to overcome this problem. The main results are summarized as follows. First, the validity of the Heckman's sample selection model is statistically significant. Male is significantly higher in both job probability and wage than female. As age increases and parents' income increases, both the probability of employment and the size of wages are higher. Finally, as the university satisfaction increases and the number of certifications acquired increased, both the probability of employment and the wage tends to increase.

Estimation of the survival function of the legislative process in Korea: based on the experiences of the 17th, 18th, and 19th National Assembly of Korea (국회 법안 검토 기간의 생존함수 추정: 제 17, 18, 19대 국회의 사례를 바탕으로)

  • Yun, Yeonggyu;Cho, Yunsoo;Jung, Hye-Young
    • The Korean Journal of Applied Statistics
    • /
    • v.32 no.4
    • /
    • pp.503-515
    • /
    • 2019
  • In this study we estimate the survival function of duration of the legislative processes in the 17th, 18th, and 19th National Assembly of Korea, and further analyze effects of the political situation variables on the legislative process. We define the termination of legislative process from a novel perspective to alleviate issues of dependency between censoring and failure in the data. We also show that the proportional hazards assumption does not hold for the data, and analyze data employing a log-normal accelerated failure time model. The policy areas of law agendas are shown to affect the speed of legislative process in different ways and legislative process tends to be prompt in times of divided governments.