• Title/Summary/Keyword: Probability Score

Search Result 295, Processing Time 0.026 seconds

Performance of APACHE IV in Medical Intensive Care Unit Patients: Comparisons with APACHE II, SAPS 3, and MPM0 III

  • Ko, Mihye;Shim, Miyoung;Lee, Sang-Min;Kim, Yujin;Yoon, Soyoung
    • Acute and Critical Care
    • /
    • v.33 no.4
    • /
    • pp.216-221
    • /
    • 2018
  • Background: In this study, we analyze the performance of the Acute Physiology and Chronic Health Evaluation (APACHE) II, APACHE IV, Simplified Acute Physiology Score (SAPS) 3, and Mortality Probability Model $(MPM)_0$ III in order to determine which system best implements data related to the severity of medical intensive care unit (ICU) patients. Methods: The present study was a retrospective investigation analyzing the discrimination and calibration of APACHE II, APACHE IV, SAPS 3, and $MPM_0$ III when used to evaluate medical ICU patients. Data were collected for 788 patients admitted to the ICU from January 1, 2015 to December 31, 2015. All patients were aged 18 years or older with ICU stays of at least 24 hours. The discrimination abilities of the three systems were evaluated using c-statistics, while calibration was evaluated by the Hosmer-Lemeshow test. A severity correction model was created using logistics regression analysis. Results: For the APACHE IV, SAPS 3, $MPM_0$ III, and APACHE II systems, the area under the receiver operating characteristic curves was 0.745 for APACHE IV, resulting in the highest discrimination among all four scoring systems. The value was 0.729 for APACHE II, 0.700 for SAP 3, and 0.670 for $MPM_0$ III. All severity scoring systems showed good calibrations: APACHE II (chi-square, 12.540; P=0.129), APACHE IV (chi-square, 6.959; P=0.541), SAPS 3 (chi-square, 9.290; P=0.318), and $MPM_0$ III (chi-square, 11.128; P=0.133). Conclusions: APACHE IV provided the best discrimination and calibration abilities and was useful for quality assessment and predicting mortality in medical ICU patients.

Enhancing Medium-Range Forecast Accuracy of Temperature and Relative Humidity over South Korea using Minimum Continuous Ranked Probability Score (CRPS) Statistical Correction Technique (연속 순위 확률 점수를 활용한 통합 앙상블 모델에 대한 기온 및 습도 후처리 모델 개발)

  • Hyejeong Bok;Junsu Kim;Yeon-Hee Kim;Eunju Cho;Seungbum Kim
    • Atmosphere
    • /
    • v.34 no.1
    • /
    • pp.23-34
    • /
    • 2024
  • The Korea Meteorological Administration has improved medium-range weather forecasts by implementing post-processing methods to minimize numerical model errors. In this study, we employ a statistical correction technique known as the minimum continuous ranked probability score (CRPS) to refine medium-range forecast guidance. This technique quantifies the similarity between the predicted values and the observed cumulative distribution function of the Unified Model Ensemble Prediction System for Global (UM EPSG). We evaluated the performance of the medium-range forecast guidance for surface air temperature and relative humidity, noting significant enhancements in seasonal bias and root mean squared error compared to observations. Notably, compared to the existing the medium-range forecast guidance, temperature forecasts exhibit 17.5% improvement in summer and 21.5% improvement in winter. Humidity forecasts also show 12% improvement in summer and 23% improvement in winter. The results indicate that utilizing the minimum CRPS for medium-range forecast guidance provide more reliable and improved performance than UM EPSG.

Improvement of Mid/Long-Term ESP Scheme Using Probabilistic Weather Forecasting (확률기상예보를 이용한 중장기 ESP기법 개선)

  • Kim, Joo-Cheol;Kim, Jeong-Kon;Lee, Sang-Jin
    • Journal of Korea Water Resources Association
    • /
    • v.44 no.10
    • /
    • pp.843-851
    • /
    • 2011
  • In hydrology, it is appropriate to use probabilistic method for forecasting mid/long term streamflow due to the uncertainty of input data. Through this study, it is expanded mid/long term forecasting system more effectively adding priory process function based on PDF-ratio method to the RRFS-ESP system for Guem River Basin. For implementing this purpose, weight is estimated using probabilistic weather forecasting information from KMA. Based on these results, ESP probability is updated per scenario. Through the estimated result per method, the average forecast score using ESP method is higher than that of naive forecasting and it confirmed that ESP method results in appropriate score for RRFS-ESP system. It is also shown that the score of ESP method applying revised inflow scenario using probabilistic weather forecasting is higher than that of ESP method. As a results, it will be improved the accuracy of forecasting using probabilistic weather forecasting.

Heart Sound-Based Cardiac Disorder Classifiers Using an SVM to Combine HMM and Murmur Scores (SVM을 이용하여 HMM과 심잡음 점수를 결합한 심음 기반 심장질환 분류기)

  • Kwak, Chul;Kwon, Oh-Wook
    • The Journal of the Acoustical Society of Korea
    • /
    • v.30 no.3
    • /
    • pp.149-157
    • /
    • 2011
  • In this paper, we propose a new cardiac disorder classification method using an support vector machine (SVM) to combine hidden Markov model (HMM) and murmur existence information. Using cepstral features and the HMM Viterbi algorithm, we segment input heart sound signals into HMM states for each cardiac disorder model and compute log-likelihood (score) for every state in the model. To exploit the temporal position characteristics of murmur signals, we divide the input signals into two subbands and compute murmur probability of every subband of each frame, and obtain the murmur score for each state by using the state segmentation information obtained from the Viterbi algorithm. With an input vector containing the HMM state scores and the murmur scores for all cardiac disorder models, SVM finally decides the cardiac disorder category. In cardiac disorder classification experimental results, the proposed method shows the relatively improvement rate of 20.4 % compared to the HMM-based classifier with the conventional cepstral features.

Estimating Average Causal Effect in Latent Class Analysis (잠재범주분석을 이용한 원인적 영향력 추론에 관한 연구)

  • Park, Gayoung;Chung, Hwan
    • The Korean Journal of Applied Statistics
    • /
    • v.27 no.7
    • /
    • pp.1077-1095
    • /
    • 2014
  • Unlike randomized trial, statistical strategies for inferring the unbiased causal relationship are required in the observational studies. Recently, new methods for the causal inference in the observational studies have been proposed such as the matching with the propensity score or the inverse probability treatment weighting. They have focused on how to control the confounders and how to evaluate the effect of the treatment on the result variable. However, these conventional methods are valid only when the treatment variable is categorical and both of the treatment and the result variables are directly observable. Research on the causal inference can be challenging in part because it may not be possible to directly observe the treatment and/or the result variable. To address this difficulty, we propose a method for estimating the average causal effect when both of the treatment and the result variables are latent. The latent class analysis has been applied to calculate the propensity score for the latent treatment variable in order to estimate the causal effect on the latent result variable. In this work, we investigate the causal effect of adolescents delinquency on their substance use using data from the 'National Longitudinal Study of Adolescent Health'.

Extracting Korean-English Parallel Sentences from Wikipedia (위키피디아로부터 한국어-영어 병렬 문장 추출)

  • Kim, Sung-Hyun;Yang, Seon;Ko, Youngjoong
    • Journal of KIISE:Software and Applications
    • /
    • v.41 no.8
    • /
    • pp.580-585
    • /
    • 2014
  • This paper conducts a variety of experiments for "the extraction of Korean parallel sentences using Wikipedia data". We refer to various methods that were previously proposed for other languages. We use two approaches. The first one is to use translation probabilities that are extracted from the existing resources such as Sejong parallel corpus, and the second one is to use dictionaries such as Wiki dictionary consisting of Wikipedia titles and MRDs (machine readable dictionaries). Experimental results show that we obtained a significant improvement in system using Wikipedia data in comparison to one using only the existing resources. We finally achieve an outstanding performance, an F1-score of 57.6%. We additionally conduct experiments using a topic model. Although this experiment shows a relatively lower performance, an F1-score of 51.6%, it is expected to be worthy of further studies.

Quantification on Dam Condition Related to Internal Erosion of an Embankment Dam and its Applicability Evaluation (필댐의 내부침식과 관련된 댐 상태의 정량화 및 적용성 평가)

  • Heo, Gun;Chung, Choong-Ki
    • Journal of the Korean Geotechnical Society
    • /
    • v.35 no.4
    • /
    • pp.5-14
    • /
    • 2019
  • The typical four conditions related to internal erosion were set from the results of the regular dam safety inspection for 17,500 dams, and a questionnaire survey was conducted for dam safety experts to quantify these four typical current dam conditions with scores between 0 and 10, respectively. In addition, we proposed 'possible score range' for each condition to minimize the decision limits for dam managers to quantify dam conditions while helping to quantify various dam conditions except 4 representative conditions. A case study based on 'quantified score' and 'possible score range' for each condition showed that this method consistently reflects the dangerousness of the dam and provides a reasonable probability of failure. This helps to overcome limitations of dam rating determination by weighted average, and it will help to evaluate dangerous dams as dangerous dams.

The Prediction of Export Credit Guarantee Accident using Machine Learning (기계학습을 이용한 수출신용보증 사고예측)

  • Cho, Jaeyoung;Joo, Jihwan;Han, Ingoo
    • Journal of Intelligence and Information Systems
    • /
    • v.27 no.1
    • /
    • pp.83-102
    • /
    • 2021
  • The government recently announced various policies for developing big-data and artificial intelligence fields to provide a great opportunity to the public with respect to disclosure of high-quality data within public institutions. KSURE(Korea Trade Insurance Corporation) is a major public institution for financial policy in Korea, and thus the company is strongly committed to backing export companies with various systems. Nevertheless, there are still fewer cases of realized business model based on big-data analyses. In this situation, this paper aims to develop a new business model which can be applied to an ex-ante prediction for the likelihood of the insurance accident of credit guarantee. We utilize internal data from KSURE which supports export companies in Korea and apply machine learning models. Then, we conduct performance comparison among the predictive models including Logistic Regression, Random Forest, XGBoost, LightGBM, and DNN(Deep Neural Network). For decades, many researchers have tried to find better models which can help to predict bankruptcy since the ex-ante prediction is crucial for corporate managers, investors, creditors, and other stakeholders. The development of the prediction for financial distress or bankruptcy was originated from Smith(1930), Fitzpatrick(1932), or Merwin(1942). One of the most famous models is the Altman's Z-score model(Altman, 1968) which was based on the multiple discriminant analysis. This model is widely used in both research and practice by this time. The author suggests the score model that utilizes five key financial ratios to predict the probability of bankruptcy in the next two years. Ohlson(1980) introduces logit model to complement some limitations of previous models. Furthermore, Elmer and Borowski(1988) develop and examine a rule-based, automated system which conducts the financial analysis of savings and loans. Since the 1980s, researchers in Korea have started to examine analyses on the prediction of financial distress or bankruptcy. Kim(1987) analyzes financial ratios and develops the prediction model. Also, Han et al.(1995, 1996, 1997, 2003, 2005, 2006) construct the prediction model using various techniques including artificial neural network. Yang(1996) introduces multiple discriminant analysis and logit model. Besides, Kim and Kim(2001) utilize artificial neural network techniques for ex-ante prediction of insolvent enterprises. After that, many scholars have been trying to predict financial distress or bankruptcy more precisely based on diverse models such as Random Forest or SVM. One major distinction of our research from the previous research is that we focus on examining the predicted probability of default for each sample case, not only on investigating the classification accuracy of each model for the entire sample. Most predictive models in this paper show that the level of the accuracy of classification is about 70% based on the entire sample. To be specific, LightGBM model shows the highest accuracy of 71.1% and Logit model indicates the lowest accuracy of 69%. However, we confirm that there are open to multiple interpretations. In the context of the business, we have to put more emphasis on efforts to minimize type 2 error which causes more harmful operating losses for the guaranty company. Thus, we also compare the classification accuracy by splitting predicted probability of the default into ten equal intervals. When we examine the classification accuracy for each interval, Logit model has the highest accuracy of 100% for 0~10% of the predicted probability of the default, however, Logit model has a relatively lower accuracy of 61.5% for 90~100% of the predicted probability of the default. On the other hand, Random Forest, XGBoost, LightGBM, and DNN indicate more desirable results since they indicate a higher level of accuracy for both 0~10% and 90~100% of the predicted probability of the default but have a lower level of accuracy around 50% of the predicted probability of the default. When it comes to the distribution of samples for each predicted probability of the default, both LightGBM and XGBoost models have a relatively large number of samples for both 0~10% and 90~100% of the predicted probability of the default. Although Random Forest model has an advantage with regard to the perspective of classification accuracy with small number of cases, LightGBM or XGBoost could become a more desirable model since they classify large number of cases into the two extreme intervals of the predicted probability of the default, even allowing for their relatively low classification accuracy. Considering the importance of type 2 error and total prediction accuracy, XGBoost and DNN show superior performance. Next, Random Forest and LightGBM show good results, but logistic regression shows the worst performance. However, each predictive model has a comparative advantage in terms of various evaluation standards. For instance, Random Forest model shows almost 100% accuracy for samples which are expected to have a high level of the probability of default. Collectively, we can construct more comprehensive ensemble models which contain multiple classification machine learning models and conduct majority voting for maximizing its overall performance.

Predictability for Heavy Rainfall over the Korean Peninsula during the Summer using TIGGE Model (TIGGE 모델을 이용한 한반도 여름철 집중호우 예측 활용에 관한 연구)

  • Hwang, Yoon-Jeong;Kim, Yeon-Hee;Chung, Kwan-Young;Chang, Dong-Eon
    • Atmosphere
    • /
    • v.22 no.3
    • /
    • pp.287-298
    • /
    • 2012
  • The predictability of heavy precipitation over the Korean Peninsula is studied using THORPEX Interactive Grand Global Ensemble (TIGGE) data. The performance of the six ensemble models is compared through the inconsistency (or jumpiness) and Root Mean Square Error (RMSE) for MSLP, T850 and H500. Grand Ensemble (GE) of the three best ensemble models (ECMWF, UKMO and CMA) with equal weight and without bias correction is consisted. The jumpiness calculated in this study indicates that the GE is more consistent than each single ensemble model. Brier Score (BS) of precipitation also shows that the GE outperforms. The GE is used for a case study of a heavy rainfall event in Korean Peninsula on 9 July 2009. The probability forecast of precipitation using 90 members of the GE and the percentage of 90 members exceeding 90 percentile in climatological Probability Density Function (PDF) of observed precipitation are calculated. As the GE is excellent in possibility of potential detection of heavy rainfall, GE is more skillful than the single ensemble model and can lead to a heavy rainfall warning in medium-range. If the performance of each single ensemble model is also improved, GE can provide better performance.

A Probabilistic Analysis for Periodicity of Real-time Tasks

  • Delgado, Raimarius;Choi, Byoung Wook
    • International Journal of Internet, Broadcasting and Communication
    • /
    • v.13 no.1
    • /
    • pp.134-142
    • /
    • 2021
  • This paper proposes a probabilistic method in analyzing timing measurements to determine the periodicity of real-time tasks. The proposed method fills a gap in existing techniques, which either concentrate on the estimation of worst-case execution times, or do not consider the stochastic behavior of the real-time scheduler. Our method is based on the Z-test statistical analysis which calculates the probability of the measured period to fall within a user-defined standard deviation limit. The distribution of the measured period should satisfy two conditions: its center (statistical mean) should be equal to the scheduled period of the real-time task, and that it should be symmetrical with most of the samples focused on the center. To ensure that these requirements are met, a data adjustment process, which omits any outliers in the expense of accuracy, is presented. Then, the Z-score of the distribution according to the user-defined deviation limit provides a probability which determines the periodicity of the real-time task. Experiments are conducted to analyze the timing measurements of real-time tasks based on real-time Linux extensions of Xenomai and RT-Preempt. The results indicate that the proposed method is able to provide easier interpretation of the periodicity of real-time tasks which are valuable especially in comparing the performance of various real-time systems.