• Title/Summary/Keyword: Validation data set

Search Result 379, Processing Time 0.027 seconds

Predicting the mortality of pneumonia patients visiting the emergency department through machine learning (기계학습모델을 통한 응급실 폐렴환자의 사망예측 모델과 기존 예측 모델의 비교)

  • Bae, Yeol;Moon, Hyung Ki;Kim, Soo Hyun
    • Journal of The Korean Society of Emergency Medicine
    • /
    • v.29 no.5
    • /
    • pp.455-464
    • /
    • 2018
  • Objective: Machine learning is not yet widely used in the medical field. Therefore, this study was conducted to compare the performance of preexisting severity prediction models and machine learning based models (random forest [RF], gradient boosting [GB]) for mortality prediction in pneumonia patients. Methods: We retrospectively collected data from patients who visited the emergency department of a tertiary training hospital in Seoul, Korea from January to March of 2015. The Pneumonia Severity Index (PSI) and Sequential Organ Failure Assessment (SOFA) scores were calculated for both groups and the area under the curve (AUC) for mortality prediction was computed. For the RF and GB models, data were divided into a test set and a validation set by the random split method. The training set was learned in RF and GB models and the AUC was obtained from the validation set. The mean AUC was compared with the other two AUCs. Results: Of the 536 investigated patients, 395 were enrolled and 41 of them died. The AUC values of PSI and SOFA scores were 0.799 (0.737-0.862) and 0.865 (0.811-0.918), respectively. The mean AUC values obtained by the RF and GB models were 0.928 (0.899-0.957) and 0.919 (0.886-0.952), respectively. There were significant differences between preexisting severity prediction models and machine learning based models (P<0.001). Conclusion: Classification through machine learning may help predict the mortality of pneumonia patients visiting the emergency department.

River Water Level Prediction Method based on LSTM Neural Network

  • Le, Xuan Hien;Lee, Giha
    • Proceedings of the Korea Water Resources Association Conference
    • /
    • 2018.05a
    • /
    • pp.147-147
    • /
    • 2018
  • In this article, we use an open source software library: TensorFlow, developed for the purposes of conducting very complex machine learning and deep neural network applications. However, the system is general enough to be applicable in a wide variety of other domains as well. The proposed model based on a deep neural network model, LSTM (Long Short-Term Memory) to predict the river water level at Okcheon Station of the Guem River without utilization of rainfall - forecast information. For LSTM modeling, the input data is hourly water level data for 15 years from 2002 to 2016 at 4 stations includes 3 upstream stations (Sutong, Hotan, and Songcheon) and the forecasting-target station (Okcheon). The data are subdivided into three purposes: a training data set, a testing data set and a validation data set. The model was formulated to predict Okcheon Station water level for many cases from 3 hours to 12 hours of lead time. Although the model does not require many input data such as climate, geography, land-use for rainfall-runoff simulation, the prediction is very stable and reliable up to 9 hours of lead time with the Nash - Sutcliffe efficiency (NSE) is higher than 0.90 and the root mean square error (RMSE) is lower than 12cm. The result indicated that the method is able to produce the river water level time series and be applicable to the practical flood forecasting instead of hydrologic modeling approaches.

  • PDF

Motion Recognition for Kinect Sensor Data Using Machine Learning Algorithm with PNF Patterns of Upper Extremities

  • Kim, Sangbin;Kim, Giwon;Kim, Junesun
    • The Journal of Korean Physical Therapy
    • /
    • v.27 no.4
    • /
    • pp.214-220
    • /
    • 2015
  • Purpose: The purpose of this study was to investigate the availability of software for rehabilitation with the Kinect sensor by presenting an efficient algorithm based on machine learning when classifying the motion data of the PNF pattern if the subjects were wearing a patient gown. Methods: The motion data of the PNF pattern for upper extremities were collected by Kinect sensor. The data were obtained from 8 normal university students without the limitation of upper extremities. The subjects, wearing a T-shirt, performed the PNF patterns, D1 and D2 flexion, extensions, 30 times; the same protocol was repeated while wearing a patient gown to compare the classification performance of algorithms. For comparison of performance, we chose four algorithms, Naive Bayes Classifier, C4.5, Multilayer Perceptron, and Hidden Markov Model. The motion data for wearing a T-shirt were used for the training set, and 10 fold cross-validation test was performed. The motion data for wearing a gown were used for the test set. Results: The results showed that all of the algorithms performed well with 10 fold cross-validation test. However, when classifying the data with a hospital gown, Hidden Markov model (HMM) was the best algorithm for classifying the motion of PNF. Conclusion: We showed that HMM is the most efficient algorithm that could handle the sequence data related to time. Thus, we suggested that the algorithm which considered the sequence of motion, such as HMM, would be selected when developing software for rehabilitation which required determining the correctness of the motion.

A Study on Stochastic Simulation Models to Internally Validate Analytical Error of a Point and a Line Segment (포인트와 라인 세그먼트의 해석적 에러 검증을 위한 확률기반 시뮬레이션 모델에 관한 연구)

  • Hong, Sung Chul;Joo, Yong Jin
    • Spatial Information Research
    • /
    • v.21 no.2
    • /
    • pp.45-54
    • /
    • 2013
  • Analytical and simulation error models have the ability to describe (or realize) error-corrupted versions of spatial data. But the different approaches for modeling positional errors require an internal validation that ascertains whether the analytical and simulation error models predict correct positional errors in a defined set of conditions. This paper presents stochastic simulation models of a point and a line segm ent to be validated w ith analytical error models, which are an error ellipse and an error band model, respectively. The simulation error models populate positional errors by the Monte Carlo simulation, according to an assumed error distribution prescribed by given parameters of a variance-covariance matrix. In the validation process, a set of positional errors by the simulation models is compared to a theoretical description by the analytical error models. Results show that the proposed simulation models realize positional uncertainties of the same spatial data according to a defined level of positional quality.

A Novel Algorithm of Joint Probability Data Association Based on Loss Function

  • Jiao, Hao;Liu, Yunxue;Yu, Hui;Li, Ke;Long, Feiyuan;Cui, Yingjie
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.15 no.7
    • /
    • pp.2339-2355
    • /
    • 2021
  • In this paper, a joint probabilistic data association algorithm based on loss function (LJPDA) is proposed so that the computation load and accuracy of the multi-target tracking algorithm can be guaranteed simultaneously. Firstly, data association is divided in to three cases based on the relationship among validation gates and the number of measurements in the overlapping area for validation gates. Also the contribution coefficient is employed for evaluating the contribution of a measurement to a target, and the loss function, which reflects the cost of the new proposed data association algorithm, is defined. Moreover, the equation set of optimal contribution coefficient is given by minimizing the loss function, and the optimal contribution coefficient can be attained by using the Newton-Raphson method. In this way, the weighted value of each target can be achieved, and the data association among measurements and tracks can be realized. Finally, we compare performances of LJPDA proposed and joint probabilistic data association (JPDA) algorithm via numerical simulations, and much attention is paid on real-time performance and estimation error. Theoretical analysis and experimental results reveal that the LJPDA algorithm proposed exhibits small estimation error and low computation complexity.

Analysis of Viscous Free Surface Flow around a Ship by a Level-set Method

  • Park, Il-Ryong;Chun, Ho-Hwan
    • Journal of Ship and Ocean Technology
    • /
    • v.6 no.2
    • /
    • pp.37-50
    • /
    • 2002
  • In the present numerical simulation of viscous free surface flow around a ship, two-fluids in-compressible Reynolds-averaged Navier-Stokes equations with the standard $\textsc{k}-\varepsilon$turbulence model are discretized on a regular grid by using a finite volume method. A local level-set method is introduced for capturing the free surface movement and the influence of the viscous layer and dynamic boundary condition of the free surface are implicitly considered. Partial differential equations in the level-set method are discretized with second order ENO scheme and explicit Euler scheme in the space and time integration, respectively. The computational results for the Series-60 model with $C_B=0.6$ show a good agreement with the experimental data, but more validation studies for commercial complicated hull forms are necessary.

A Study on Robust Identification Based on the Validation Evaluation of Model (모델의 타당성 평가에 기초한 로바스트 동정에 관한 연구)

  • Lee, D.C.
    • Journal of Power System Engineering
    • /
    • v.4 no.3
    • /
    • pp.72-80
    • /
    • 2000
  • In order to design a stable robust controller, nominal model, and the upper bound about the uncertainty which is the error of the model are needed. The problem to estimate the nominal model of controlled system and the upper bound of uncertainty at the same time is called robust identification. When the nominal model of controlled system and the upper bound of uncertainty in relation to robust identification are given, the evaluation of the validity of the model and the upper bound makes it possible to distinguish whether there is a model which explains observation data including disturbance among the model set. This paper suggests a method to identity the uncertainty which removes disturbance and expounds observation data by giving a probable postulation and plural data set to disturbance. It also examines the suggested method through a numerical computation simulation and validates its effectiveness.

  • PDF

A Study on Robust Identification Based on the Validation Evaluation of Model (모델의 타당성 평가에 기초한 로바스트 동정에 관한 연구)

  • Lee, Dong-Cheol;Chung, Hyung-Hwan;Bae, Jong-Il
    • Proceedings of the KIEE Conference
    • /
    • 2000.07d
    • /
    • pp.2690-2692
    • /
    • 2000
  • In order to design a stable robust controller, nominal model, and the upper bound about the uncertainty which is the error of the model are needed. The problem to estimate the nominal model of controlled system and the upper bound of uncertainty at the same time is called robust identifcation. When the nominal model of controlled system and the upper bound of uncertainty in relation to robust identifcation are given, the evaluation of the validity of the model and the upper bound makes it possible to distinguish whether there is a model which explains observation data including disturbance among the model set. This paper suggests a method to identify the uncertainty which removes disturbance and expounds observation data by giving a probable postulation and plural data set to disturbance. It also examines the suggested method through a numerical computation simulation and validates its effectiveness.

  • PDF

Cold sensitivity classification using facial image based on convolutional neural network

  • lkoo Ahn;Younghwa Baek;Kwang-Ho Bae;Bok-Nam Seo;Kyoungsik Jung;Siwoo Lee
    • The Journal of Korean Medicine
    • /
    • v.44 no.4
    • /
    • pp.136-149
    • /
    • 2023
  • Objectives: Facial diagnosis is an important part of clinical diagnosis in traditional East Asian Medicine. In this paper, we proposed a model to quantitatively classify cold sensitivity using a fully automated facial image analysis system. Methods: We investigated cold sensitivity in 452 subjects. Cold sensitivity was determined using a questionnaire and the Cold Pattern Score (CPS) was used for analysis. Subjects with a CPS score below the first quartile (low CPS group) belonged to the cold non-sensitivity group, and subjects with a CPS score above the third quartile (high CPS group) belonged to the cold sensitivity group. After splitting the facial images into train/validation/test sets, the train and validation set were input into a convolutional neural network to learn the model, and then the classification accuracy was calculated for the test set. Results: The classification accuracy of the low CPS group and high CPS group using facial images in all subjects was 76.17%. The classification accuracy by sex was 69.91% for female and 62.86% for male. It is presumed that the deep learning model used facial color or facial shape to classify the low CPS group and the high CPS group, but it is difficult to specifically determine which feature was more important. Conclusions: The experimental results of this study showed that the low CPS group and the high CPS group can be classified with a modest level of accuracy using only facial images. There was a need to develop more advanced models to increase classification accuracy.

The Integrated Methodology of Rough Set Theory and Artificial Neural Network for Business Failure Prediction (도산 예측을 위한 러프집합이론과 인공신경망 통합방법론)

  • Kim, Chang-Yun;Ahn, Byeong-Seok;Cho, Sung-Sik;Kim, Soung-Hie
    • Asia pacific journal of information systems
    • /
    • v.9 no.4
    • /
    • pp.23-40
    • /
    • 1999
  • This paper proposes a hybrid intelligent system that predicts the failure of firms based on the past financial performance data, combining neural network and rough set approach, We can get reduced information table, which implies that the number of evaluation criteria such as financial ratios and qualitative variables and objects (i.e., firms) is reduced with no information loss through rough set approach. And then, this reduced information is used to develop classification rules and train neural network to infer appropriate parameters. Through the reduction of information table, it is expected that the performance of the neural network improve. The rules developed by rough sets show the best prediction accuracy if a case does match any of the rules. The rationale of our hybrid system is using rules developed by rough sets for an object that matches any of the rules and neural network for one that does not match any of them. The effectiveness of our methodology was verified by experiments comparing traditional discriminant analysis and neural network approach with our hybrid approach. For the experiment, the financial data of 2,400 Korean firms during the period 1994-1996 were selected, and for the validation, k-fold validation was used.

  • PDF