• Title/Summary/Keyword: Data Validation

Search Result 3,256, Processing Time 0.032 seconds

Validation Technique of Trace-Driven Simulation Model Using Weighted F-measure (가중 F 척도를 이용한 Trace-Driven 시뮬레이션 모델의 검증 방법)

  • HwangBo, Hoon;Cheon, Hyeon-Jae;Lee, Hong-Chul
    • Journal of the Korea Society for Simulation
    • /
    • v.18 no.4
    • /
    • pp.185-195
    • /
    • 2009
  • As most systems get more complicated, system analysis using simulation has been taken notice of. One of the core parts of simulation analysis is validation of a simulation model, and we can identify how well the simulation model represents the real system with this validation process. The difference between input data of two systems has an effect on the comparison between a simulation model and a real system at validation stage, and the result with such difference is not enough to ensure high credibility of the model. Accordingly, in this paper, we construct a model based on Trace-driven simulation which uses identical input data with the real system. On the other hand, to validate a model by each class, not by an unique statistic, we validate the model using a metric transformed from F-measure which estimates performance of a classifier in data mining field. Finally, this procedure enables precise validation process of a model, and it helps modification by offering feedback at the validation phase.

Validation Method of Simulation Model Using Wavelet Transform (웨이블릿 변환을 이용한 시뮬레이션 모델 검증 방법)

  • Shin, Sang-Mi;Kim, Youn-Jin;Lee, Hong-Chul
    • Journal of the Korea Society for Simulation
    • /
    • v.19 no.2
    • /
    • pp.127-135
    • /
    • 2010
  • The validation of a simulation model is a key to demonstrate that the simulation model is reliable. However, among various validation methods have been introduced, it is very poor to research the specific techniques for the time series data. Therefore, this paper suggests the methodology to verify the simulation using the time series data by Wavelet Transform, Power Spectrum and Coherence. This method performs 2 steps as followed. Firstly, we get spectrum using the Wavelet transform available for non-periodic signal separation. Secondly, we compare 2 patterns of output data from simulation model and actual system by Coherence Analysis. As a result of comparing it with other validation techniques, the suggested way can judge simulation model accuracy more clearly. By this way, we can make it possible to perform the simulation validation test under various situations using detailed sectional validation method, which has been impossible using a single statistics for the whole model.

Validation of Quality Control Algorithms for Temperature Data of the Republic of Korea (한국의 기온자료 품질관리 알고리즘의 검증)

  • Park, Changyong;Choi, Youngeun
    • Atmosphere
    • /
    • v.22 no.3
    • /
    • pp.299-307
    • /
    • 2012
  • This study is aimed to validate errors for detected suspicious temperature data using various quality control procedures for 61 weather stations in the Republic of Korea. The quality control algorithms for temperature data consist of four main procedures (high-low extreme check, internal consistency check, temporal outlier check, and spatial outlier check). Errors of detected suspicious temperature data are judged by examining temperature data of nearby stations, surface weather charts, hourly temperature data, daily precipitation, and daily maximum wind direction. The number of detected errors in internal consistency check and spatial outlier check showed 4 days (3 stations) and 7 days (5 stations), respectively. Effective and objective methods for validation errors through this study will help to reduce manpower and time for conduct of quality management for temperature data.

Advances in the Development and Validation of Test Methods in the United States

  • Casey, Warren M.
    • Toxicological Research
    • /
    • v.32 no.1
    • /
    • pp.9-14
    • /
    • 2016
  • The National Toxicology Program Interagency Center for the Evaluation of Alternative Toxicological Methods (NICEATM) provides validation support for US Federal agencies and the US Tox21 interagency consortium, an interagency collaboration that is using high throughput screening (HTS) and other advanced approaches to better understand and predict chemical hazards to humans and the environment. The use of HTS data from assays relevant to the estrogen receptor signaling data pathway is used as an example of how HTS data can be combined with computational modeling to meet the needs of US agencies. As brief summary of US efforts in the areas of biologics testing, acute toxicity, and skin sensitization will also be provided.

A Study on HVAC Parameter Monitoring System (Regarding Computer Validation) (HVAC 파라미터 모니터링 시스템에 대한 고찰 (Computer Validation 중심으로))

  • Kim, Jong-Gu
    • Proceedings of the SAREK Conference
    • /
    • 2008.06a
    • /
    • pp.90-95
    • /
    • 2008
  • This article presents practical advice regarding the implementation and management of an impeccable Building Management System. The BMS was introduced to the series of computerized systems including manufacturing, storage, distribution, and quality control. Recently revised GMP regulation is requesting an improvement in drug product quality regulatory system by computer system validation. Quality is critical to guarantee the efficacy and the safety of drugs and is approved in the evaluation process after the audit trail application. HVAC parameter monitoring system will record the identity of operators entering or confirming critical data. Authority to amend entered data should be restricted to nominated persons. Any alteration to an entry of critical data should be authorized in advance and recorded with the reason for the change.

  • PDF

Predicting the popularity of TV-show through text mining of tweets: A Drama Case in South Korea

  • Kim, Do Yeon;Kim, Yoosin;Choi, Sang Hyun
    • Journal of Internet Computing and Services
    • /
    • v.17 no.5
    • /
    • pp.131-139
    • /
    • 2016
  • This paper presents a workflow validation method for data-intensive graphical workflow models using real-time workflow tracing mode on data-intensive workflow designer. In order to model and validate workflows, we try to divide as modes have editable mode and tracing mode on data-intensive workflow designer. We could design data-intensive workflow using drag and drop in editable-mode, otherwise we could not design but view and trace workflow model in tracing mode. We would like to focus on tracing-mode for workflow validation, and describe how to use workflow tracing on data-intensive workflow model designer. Especially, it is support data centered operation about control logics and exchange variables on workflow runtime for workflow tracing.

Development of Performance Analysis Methodology for Nuclear Power Plant Turbine Cycle Using Validation Model of Performance Measurements (원전 터빈사이클 성능 데이터의 검증 모델에 의한 성능분석 기법의 개발)

  • Kim, Seong-Geun;Choe, Gwang-Hui
    • Transactions of the Korean Society of Mechanical Engineers B
    • /
    • v.24 no.12
    • /
    • pp.1625-1634
    • /
    • 2000
  • Verification of measurements is required for precise evaluation of turbine cycle performance in nuclear power plant. We assumed that initial acceptance data and design data of the plant could provide correlation information between performance data. The data can be used as sample sets for the correct estimation model of measurement value. The modeling was done practically by using regression model based on plant design data, plant acceptance data and verified plant performance data of domestic nuclear power plant. We can construct more robust performance analysis system for an operation nuclear power plant with this validation scheme.

A Study of Optimal Ratio of Data Partition for Neuro-Fuzzy-Based Software Reliability Prediction (뉴로-퍼지 소프트웨어 신뢰성 예측에 대한 최적의 데이터 분할비율에 관한 연구)

  • Lee, Sang-Un
    • The KIPS Transactions:PartD
    • /
    • v.8D no.2
    • /
    • pp.175-180
    • /
    • 2001
  • This paper presents the optimal fraction of validation set to obtain a prediction accuracy of software failure count or failure time in the future by a neuro-fuzzy system. Given a fixed amount of training data, the most popular effective approach to avoiding underfitting and overfitting is early stopping, and hence getting optimal generalization. But there is unresolved practical issues : How many data do you assign to the training and validation set\ulcorner Rules of thumb abound, the solution is acquired by trial-and-error and we spend long time in this method. For the sake of optimal fraction of validation set, the variant specific fraction for the validation set be provided. It shows that minimal fraction of the validation data set is sufficient to achieve good next-step prediction. This result can be considered as a practical guideline in a prediction of software reliability by neuro-fuzzy system.

  • PDF

LS-SVM for large data sets

  • Park, Hongrak;Hwang, Hyungtae;Kim, Byungju
    • Journal of the Korean Data and Information Science Society
    • /
    • v.27 no.2
    • /
    • pp.549-557
    • /
    • 2016
  • In this paper we propose multiclassification method for large data sets by ensembling least squares support vector machines (LS-SVM) with principal components instead of raw input vector. We use the revised one-vs-all method for multiclassification, which is one of voting scheme based on combining several binary classifications. The revised one-vs-all method is performed by using the hat matrix of LS-SVM ensemble, which is obtained by ensembling LS-SVMs trained using each random sample from the whole large training data. The leave-one-out cross validation (CV) function is used for the optimal values of hyper-parameters which affect the performance of multiclass LS-SVM ensemble. We present the generalized cross validation function to reduce computational burden of leave-one-out CV functions. Experimental results from real data sets are then obtained to illustrate the performance of the proposed multiclass LS-SVM ensemble.

STATUS AND PERSPECTIVE OF NUCLEAR DATA PRODUCTION, EVALUATION AND VALIDATION

  • TRKOV A.
    • Nuclear Engineering and Technology
    • /
    • v.37 no.1
    • /
    • pp.11-24
    • /
    • 2005
  • A very important feature in the development of nuclear technology has been and will continue to be the flow of information from nuclear data production to the various applications fields in nuclear technology. Both, nuclear data and this communications flow are defined in this paper. Nuclear data result from specific technical activities including their production, evaluation, processing, verification, validation and applications. These activities are described, focusing on nuclear reactor calculations. Mathematical definitions of different types of nuclear data are introduced, and international forums involved in nuclear data activities are listed. Electronic links to various sources of information available on the web are specified, whenever possible.