• Title/Summary/Keyword: research data quality management model

Search Result 840, Processing Time 0.028 seconds

Proposal of Process Model for Research Data Quality Management (연구데이터 품질관리를 위한 프로세스 모델 제안)

  • Na-eun Han
    • Journal of the Korean Society for information Management
    • /
    • v.40 no.1
    • /
    • pp.51-71
    • /
    • 2023
  • This study analyzed the government data quality management model, big data quality management model, and data lifecycle model for research data management, and analyzed the components common to each data quality management model. Those data quality management models are designed and proposed according to the lifecycle or based on the PDCA model according to the characteristics of target data, which is the object that performs quality management. And commonly, the components of planning, collection and construction, operation and utilization, and preservation and disposal are included. Based on this, the study proposed a process model for research data quality management, in particular, the research data quality management to be performed in a series of processes from collecting to servicing on a research data platform that provides services using research data as target data was discussed in the stages of planning, construction and operation, and utilization. This study has significance in providing knowledge based for research data quality management implementation methods.

A Data Quality Management Maturity Model

  • Ryu, Kyung-Seok;Park, Joo-Seok;Park, Jae-Hong
    • ETRI Journal
    • /
    • v.28 no.2
    • /
    • pp.191-204
    • /
    • 2006
  • Many previous studies of data quality have focused on the realization and evaluation of both data value quality and data service quality. These studies revealed that poor data value quality and poor data service quality were caused by poor data structure. In this study we focus on metadata management, namely, data structure quality and introduce the data quality management maturity model as a preferred maturity model. We empirically show that data quality improves as data management matures.

  • PDF

Analysis of IT Service Quality Elements Using Text Sentiment Analysis (텍스트 감정분석을 이용한 IT 서비스 품질요소 분석)

  • Kim, Hong Sam;Kim, Chong Su
    • Journal of Korean Society of Industrial and Systems Engineering
    • /
    • v.43 no.4
    • /
    • pp.33-40
    • /
    • 2020
  • In order to satisfy customers, it is important to identify the quality elements that affect customers' satisfaction. The Kano model has been widely used in identifying multi-dimensional quality attributes in this purpose. However, the model suffers from various shortcomings and limitations, especially those related to survey practices such as the data amount, reply attitude and cost. In this research, a model based on the text sentiment analysis is proposed, which aims to substitute the survey-based data gathering process of Kano models with sentiment analysis. In this model, from the set of opinion text, quality elements for the research are extracted using the morpheme analysis. The opinions' polarity attributes are evaluated using text sentiment analysis, and those polarity text items are transformed into equivalent Kano survey questions. Replies for the transformed survey questions are generated based on the total score of the original data. Then, the question-reply set is analyzed using both the original Kano evaluation method and the satisfaction index method. The proposed research model has been tested using a large amount of data of public IT service project evaluations. The result shows that it can replace the existing practice and it promises advantages in terms of quality and cost of data gathering. The authors hope that the proposed model of this research may serve as a new quality analysis model for a wide range of areas.

Quality Design Support System based on Data Mining Approach (데이터 마이닝 기반의 품질설계지원시스템)

  • 지원철
    • Journal of the Korean Operations Research and Management Science Society
    • /
    • v.28 no.3
    • /
    • pp.31-47
    • /
    • 2003
  • Quality design in practice highly depends on human designer's intuition and past experiences due to lack of formal knowledge about the relationship among 10 variables. This paper represents an data mining approach for developing quality design support system that integrates Case Based Reasoning (CBR) and Artificial Neural Networks (ANN) to effectively support all the steps in quality design process. CBR stores design cases in a systematic way and retrieve them quickly and accurately. ANN predicts the resulting quality attributes of design alternatives that are generated from CBR's adaptation process. When the predicted attributes fail to meet the target values, quality design simulation starts to further adapt the alternatives to the customer's new orders. To implement the quality design simulation, this paper suggests (1) the data screening method based on ξ-$\delta$ Ball to obtain the robust ANN models from the large production data bases, (2) the procedure of quality design simulation using ANN and (3) model management system that helps users find the appropriate one from the ANN model base. The integration of CBR and ANN provides quality design engineers the way that produces consistent and reliable design solutions in the remarkably reduced time.

A Resource Allocation Model for Data QC Activities Using Cost of Quality (품질코스트를 이용한 데이터 QC 활동의 자원할당 모형 연구)

  • Lee, Sang-Cheol;Shin, Wan-Seon
    • IE interfaces
    • /
    • v.24 no.2
    • /
    • pp.128-138
    • /
    • 2011
  • This research proposes a resource allocation model of Data QC (Quality Control) activities using COQ (Cost of Quality). The model has been developed based on a series of research efforts such as COQ classifications, weight determination of Data QC activities, and an aggregation approach between COQ and Data QC activities. In the first stage of this research, COQ was divided into the four typical classifications (prevention costs, appraisal costs, internal failure costs and external failure costs) through the opinions from five professionals in Data QC. In the second stage, the weights of Data QC activities were elicited from the field professionals. An aggregation model between COQ and Data QC activities has been then proposed to help the practitioners make a resource allocation strategy. DEA (Data Envelopment Analysis) was utilized for locating efficient decision points. The proposed resource allocation model has been validated using the case of Korea national defense information system. This research is unique in that it applies the concept of COQ to the data management for the first time and that it demonstrates a possible contribution to a real world case for budget allocation of national defense information.

Developing a Web-based System for Computing Pre-Harvest Residue Limits (PHRLs)

  • Chang, Han Sub;Bae, Hey Ree;Son, Young Bae;Song, In Ho;Lee, Cheol Ho;Choi, Nam Geun;Cho, Kyoung Kyu;Lee, Young Gu
    • Agribusiness and Information Management
    • /
    • v.3 no.1
    • /
    • pp.11-22
    • /
    • 2011
  • This study describes the development of a web-based system that collects all data generated in the research conducted to set pre-harvest residue limits (PHRLs) for agricultural product safety control. These data, including concentrations of pesticide residues, limit of detection, limit of quantitation, recoveries, weather charts, and growth rates, are incorporated into a database, a regression analysis of the data is performed using statistical techniques, and the PHRL for an agricultural product is automatically computed. The development and establishment of this system increased the efficiency and improved the reliability of the research in this area by standardizing the data and maintaining its accuracy without temporal or spatial limitations. The system permits automatic computation of the PHRL and a quick review of the goodness of fit of the regression model. By building and analyzing a database, it also allows data accumulated over the last 10 years to be utilized.

  • PDF

A study on the data quality management evaluation model (데이터 품질관리 평가 모델에 관한 연구)

  • Kim, Hyung-Sub
    • Journal of the Korea Convergence Society
    • /
    • v.11 no.7
    • /
    • pp.217-222
    • /
    • 2020
  • This study is about the data quality management evaluation model. As the information and communication technology is advanced and the importance of storage and management begins to increase, the guam feeling for data is increasing. In particular, interest in the fourth industrial revolution and artificial intelligence has been increasing recently. Data is important in the fourth industrial revolution and the era of artificial intelligence. In the 21st century, data will likely play a role as a new crude oil. It can be said that the management of the quality of this data is very important. However, research is being conducted at a practical level, but research at an academic level is insufficient. Therefore, this study examined factors affecting data quality management for experts and suggested implications. As a result of the analysis, there was a difference in the importance of data quality management.

The Study on Methods for Setting of Water Quality Goal and Estimation of Allocation Loads on TMDL System Using a Dynamic Water Quality Model (동적모델을 이용한 수질오염총량제 목표수질 설정 및 할당부하량 산정방안 연구)

  • Kim, Eun-Jung;Park, Bae-Kyung;Shin, Dong-Suk;Kim, Yong-Seok;Rhew, Doug-Hee
    • Journal of Korean Society of Environmental Engineers
    • /
    • v.36 no.9
    • /
    • pp.629-640
    • /
    • 2014
  • In this study, the applicability of dynamic water quality model to TMDL system was examined, methods for setting of water quality goal and estimation of allocation loads were suggested and results of applying these methods were analyzed. HSPF model was applied for Miho stream basin as a dynamic water quality model. The model was calibrated using measurement data obtained in 2009~2010 and showed satisfactory performance in predicting daily variations of flow rate and BOD concentration. Methods for TMDL application were categorized into 3 cases; water quality management (1) considering low flow condition(Case I), (2) considering entire period of the year (Case II) and (3) considering the worst water quality condition (Case III). BOD water quality goals at the end of Miho stream watershed increased in the order of Case IICase I>Case III. If further researches on base precipitation and method for model input of nonpoint source pollutant were carried out, water quality could be managed more reasonably and scientifically by applying dynamic water quality model to TMDL. The result of this study is expected to be used as primary data for TMDL using dynamic water quality model.

A Proposal of Evaluation of Large Language Models Built Based on Research Data (연구데이터 관점에서 본 거대언어모델 품질 평가 기준 제언)

  • Na-eun Han;Sujeong Seo;Jung-ho Um
    • Journal of the Korean Society for information Management
    • /
    • v.40 no.3
    • /
    • pp.77-98
    • /
    • 2023
  • Large Language Models (LLMs) are becoming the major trend in the natural language processing field. These models were built based on research data, but information such as types, limitations, and risks of using research data are unknown. This research would present how to analyze and evaluate the LLMs that were built with research data: LLaMA or LLaMA base models such as Alpaca of Stanford, Vicuna of the large model systems organization, and ChatGPT from OpenAI from the perspective of research data. This quality evaluation focuses on the validity, functionality, and reliability of Data Quality Management (DQM). Furthermore, we adopted the Holistic Evaluation of Language Models (HELM) to understand its evaluation criteria and then discussed its limitations. This study presents quality evaluation criteria for LLMs using research data and future development directions.

Applying Formal Methods to Modeling and Analysis of Real-time Data Streams

  • Kapitanova, Krasimira;Wei, Yuan;Kang, Woo-Chul;Son, Sang-H.
    • Journal of Computing Science and Engineering
    • /
    • v.5 no.1
    • /
    • pp.85-110
    • /
    • 2011
  • Achieving situation awareness is especially challenging for real-time data stream applications because they i) operate on continuous unbounded streams of data, and ii) have inherent realtime requirements. In this paper we showed how formal data stream modeling and analysis can be used to better understand stream behavior, evaluate query costs, and improve application performance. We used MEDAL, a formal specification language based on Petri nets, to model the data stream queries and the quality-of-service management mechanisms of RT-STREAM, a prototype system for data stream management. MEDAL's ability to combine query logic and data admission control in one model allows us to design a single comprehensive model of the system. This model can be used to perform a large set of analyses to help improve the application's performance and quality of service.