• Title/Summary/Keyword: dataset

Search Result 3,827, Processing Time 0.028 seconds

Design of Data-centroid Radial Basis Function Neural Network with Extended Polynomial Type and Its Optimization (데이터 중심 다항식 확장형 RBF 신경회로망의 설계 및 최적화)

  • Oh, Sung-Kwun;Kim, Young-Hoon;Park, Ho-Sung;Kim, Jeong-Tae
    • The Transactions of The Korean Institute of Electrical Engineers
    • /
    • v.60 no.3
    • /
    • pp.639-647
    • /
    • 2011
  • In this paper, we introduce a design methodology of data-centroid Radial Basis Function neural networks with extended polynomial function. The two underlying design mechanisms of such networks involve K-means clustering method and Particle Swarm Optimization(PSO). The proposed algorithm is based on K-means clustering method for efficient processing of data and the optimization of model was carried out using PSO. In this paper, as the connection weight of RBF neural networks, we are able to use four types of polynomials such as simplified, linear, quadratic, and modified quadratic. Using K-means clustering, the center values of Gaussian function as activation function are selected. And the PSO-based RBF neural networks results in a structurally optimized structure and comes with a higher level of flexibility than the one encountered in the conventional RBF neural networks. The PSO-based design procedure being applied at each node of RBF neural networks leads to the selection of preferred parameters with specific local characteristics (such as the number of input variables, a specific set of input variables, and the distribution constant value in activation function) available within the RBF neural networks. To evaluate the performance of the proposed data-centroid RBF neural network with extended polynomial function, the model is experimented with using the nonlinear process data(2-Dimensional synthetic data and Mackey-Glass time series process data) and the Machine Learning dataset(NOx emission process data in gas turbine plant, Automobile Miles per Gallon(MPG) data, and Boston housing data). For the characteristic analysis of the given entire dataset with non-linearity as well as the efficient construction and evaluation of the dynamic network model, the partition of the given entire dataset distinguishes between two cases of Division I(training dataset and testing dataset) and Division II(training dataset, validation dataset, and testing dataset). A comparative analysis shows that the proposed RBF neural networks produces model with higher accuracy as well as more superb predictive capability than other intelligent models presented previously.

The Validation Study of Normality Distribution of Aquatic Toxicity Data for Statistical Analysis (수생태 독성자료의 정규성 분포 특성 확인을 통해 통계분석 시 분포 특성 적용에 대한 타당성 확인 연구)

  • OK, Seung-yeop;Moon, Hyo-Bang;Ra, Jin-Sung
    • Journal of Environmental Health Sciences
    • /
    • v.45 no.2
    • /
    • pp.192-202
    • /
    • 2019
  • Objectives: According to the central limit theorem, the samples in population might be considered to follow normal distribution if a large number of samples are available. Once we assume that toxicity dataset follow normal distribution, we can treat and process data statistically to calculate genus or species mean value with standard deviation. However, little is known and only limited studies are conducted to investigate whether toxicity dataset follows normal distribution or not. Therefore, the purpose of study is to evaluate the generally accepted normality hypothesis of aquatic toxicity dataset Methods: We selected the 8 chemicals, which consist of 4 organic and 4 inorganic chemical compounds considering data availability for the development of species sensitivity distribution. Toxicity data were collected at the US EPA ECOTOX Knowledgebase by simple search with target chemicals. Toxicity data were re-arranged to a proper format based on the endpoint and test duration, where we conducted normality test according to the Shapiro-Wilk test. Also we investigated the degree of normality by simple log transformation of toxicity data Results: Despite of the central limit theorem, only one large dataset (n>25) follow normal distribution out of 25 large dataset. By log transforming, more 7 large dataset show normality. As a result of normality test on small dataset (n<25), log transformation of toxicity value generally increases normality. Both organic and inorganic chemicals show normality growth for 26 species and 30 species, respectively. Those 56 species shows normality growth by log transformation in the taxonomic groups such as amphibian (1), crustacean (21), fish (22), insect (5), rotifer (2), and worm (5). In contrast, mollusca shows normality decrease at 1 species out of 23 that originally show normality. Conclusions: The normality of large toxicity dataset was not always satisfactory to the central limit theorem. Normality of those data could be improved through log transformation. Therefore, care should be taken when using toxicity data to induce, for example, mean value for risk assessment.

A Study on Data Adjustment and Quality Enhancement Method for Public Administrative Dataset Records in the Transfer Process-Based on the Experiences of Datawarehouses' ETT (행정정보 데이터세트 기록 이관 시 데이터 보정 및 품질 개선 방법 연구 - 데이터웨어하우스 ETT 경험을 기반으로)

  • Yim, Jin-Hee;Cho, Eun-Hee
    • The Korean Journal of Archival Studies
    • /
    • no.25
    • /
    • pp.91-129
    • /
    • 2010
  • As it grows more heavily reliant on information system, researchers seek for various ways to manage and utilize of dataset records which is accumulated in public information system. It might be needed to adjust date and enhance the quality of public administrative dataset records during transferring to archive system or sharing server. The purpose of this paper is presenting data adjustment and quality enhancement methods for public administrative dataset records, and it refers to ETT procedure and method of construction of datawarehouses. It suggests seven typical examples and processing method of data adjustment and quality enhancement, which are (1) verification of quantity and data domain (2) code conversion for a consistent code value (3) making component with combinded information (4) making a decision of precision of date data (5) standardization of data (6) comment information about code value (7) capturing of metadata. It should be reviewed during dataset record transfer. This paper made Data adjustment and quality enhancement requirements for dataset record transfer, and it could be used as data quality requirement of administrative information system which produces dataset.

A Study on Data Quality Evaluation of Administrative Information Dataset (행정정보데이터세트의 데이터 품질평가 연구)

  • Song, Chiho;Yim, Jinhee
    • The Korean Journal of Archival Studies
    • /
    • no.71
    • /
    • pp.237-272
    • /
    • 2022
  • In 2019, the pilot project to establish a record management system for administrative information datasets started in earnest under the leadership of the National Archives. Based on the results of the three-year project by 2021, the improved administrative information dataset management plan will be reflected in public records-related laws and guidelines. Through this, the administrative information dataset becomes the target of full-scale public record management. Although public records have been converted to electronic documents and even the datasets of administrative information systems have been included in full-scale public records management, research on the quality requirements of data itself as raw data constituting records is still lacking. If data quality is not guaranteed, all four properties of records will be threatened in the dataset, which is a structure of data and an aggregate of records. Moreover, if the reliability of the quality of the data of the administrative information system built by reflecting the various needs of the working departments of the institution without considering the standards of the standard records management system is insufficient, the reliability of the public records itself can not be secured. This study is based on the administrative information dataset management plan presented in the "Administrative Information Dataset Recorded Information Service and Utilization Model Study" conducted by the National Archives of Korea in 2021. A study was conducted. By referring to various data, especially public data-related policies and guides, which are being promoted across the government, we would like to derive quality evaluation requirements in terms of records management and present specific indicators. Through this, it is expected that it will be helpful for record management of administrative information dataset which will be in full swing in the future.

Building Dataset of Sensor-only Facilities for Autonomous Cooperative Driving

  • Hyung Lee;Chulwoo Park;Handong Lee;Junhyuk Lee
    • Journal of the Korea Society of Computer and Information
    • /
    • v.29 no.1
    • /
    • pp.21-30
    • /
    • 2024
  • In this paper, we propose a method to build a sample dataset of the features of eight sensor-only facilities built as infrastructure for autonomous cooperative driving. The feature extracted from point cloud data acquired by LiDAR and build them into the sample dataset for recognizing the facilities. In order to build the dataset, eight sensor-only facilities with high-brightness reflector sheets and a sensor acquisition system were developed. To extract the features of facilities located within a certain measurement distance from the acquired point cloud data, a cylindrical projection method was applied to the extracted points after applying DBSCAN method for points and then a modified OTSU method for reflected intensity. Coordinates of 3D points, projected coordinates of 2D, and reflection intensity were set as the features of the facility, and the dataset was built along with labels. In order to check the effectiveness of the facility dataset built based on LiDAR data, a common CNN model was selected and tested after training, showing an accuracy of about 90% or more, confirming the possibility of facility recognition. Through continuous experiments, we will improve the feature extraction algorithm for building the proposed dataset and improve its performance, and develop a dedicated model for recognizing sensor-only facilities for autonomous cooperative driving.

From TMJ to 3D Digital Smile Design with Virtual Patient Dataset for diagnosis and treatment planning (가상환자 데이터세트를 기반으로 악관절과 심미를 고려한 진단 및 치료계획 수립)

  • Lee, Soo Young;Kang, Dong Huy;Lee, Doyun;Kim, Heechul
    • Journal of the Korean Academy of Esthetic Dentistry
    • /
    • v.30 no.2
    • /
    • pp.71-90
    • /
    • 2021
  • The virtual patient dataset is a collection of diagnostic data from various sources acquired from a single patient into a coordinate system of three-dimensional visualization. Virtual patient dataset makes it possible to establish a treatment plan, simulate various treatment procedures, and create a treatment planning delivery device. Clinicians can design and simulate a patient's smile on the virtual patient dataset and select the optimal result from the diagnostic process. The selected treatment plan can be delivered identically to the patient using manufacturing techniques such as 3D printing, milling, and injection molding. The delivery of this treatment plan can be linked to the final prosthesis through mockup confirmation through provisional restoration fabrication and delivery in the patient's mouth. In this way, if the diagnostic data superimposition and processing accuracy during the manufacturing process are guaranteed, 3D digital smile design simulated in 3D visualization can be accurately delivered to the real patient. As a clinical application method of the virtual patient dataset, we suggest a decision-making method that can exclude occlusal adjustment treatment from the treatment plan through the digital occlusal pressure analysis. A comparative analysis of whole-body scans before and after temporomandibular joint treatment was suggested for adolescent idiopathic scoliosis patients with temporomandibular joint disease. Occlusal plane and smile aesthetic analysis based on the virtual patient dataset was presented when treating patients with complete dentures.

A Study on Dataset Generation Method for Korean Language Information Extraction from Generative Large Language Model and Prompt Engineering (생성형 대규모 언어 모델과 프롬프트 엔지니어링을 통한 한국어 텍스트 기반 정보 추출 데이터셋 구축 방법)

  • Jeong Young Sang;Ji Seung Hyun;Kwon Da Rong Sae
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.12 no.11
    • /
    • pp.481-492
    • /
    • 2023
  • This study explores how to build a Korean dataset to extract information from text using generative large language models. In modern society, mixed information circulates rapidly, and effectively categorizing and extracting it is crucial to the decision-making process. However, there is still a lack of Korean datasets for training. To overcome this, this study attempts to extract information using text-based zero-shot learning using a generative large language model to build a purposeful Korean dataset. In this study, the language model is instructed to output the desired result through prompt engineering in the form of "system"-"instruction"-"source input"-"output format", and the dataset is built by utilizing the in-context learning characteristics of the language model through input sentences. We validate our approach by comparing the generated dataset with the existing benchmark dataset, and achieve 25.47% higher performance compared to the KLUE-RoBERTa-large model for the relation information extraction task. The results of this study are expected to contribute to AI research by showing the feasibility of extracting knowledge elements from Korean text. Furthermore, this methodology can be utilized for various fields and purposes, and has potential for building various Korean datasets.

Statistical Blade Angular Velocity Information-based Wind Turbine Fault Diagnosis Monitoring System (블레이드 각속도 통계 정보 기반 풍력 발전기 고장 진단 모니터링 시스템)

  • Kim, Byoungjin;Kang, Suk-Ju;Park, Joon-Young
    • KEPCO Journal on Electric Power and Energy
    • /
    • v.2 no.4
    • /
    • pp.619-625
    • /
    • 2016
  • In this paper, we propose a new fault diagnosis monitoring system using gyro sensor-based angular velocity calculation for blades of the wind turbine system. First, the proposed system generates the angular velocity dataset for the rotation speed of the normal blade. Using the dataset, we estimate and evaluate the state of blades for the wind turbine by comparing the current state with the pre-calculated normal state. In the experimental results, the angular velocity of the normal state was higher than $360^{\circ}/s$ while that of the damaged blades was lower than $360^{\circ}/s$ and the standard deviation of the angular velocity was significantly increased.

Construction of a Video Dataset for Face Tracking Benchmarking Using a Ground Truth Generation Tool

  • Do, Luu Ngoc;Yang, Hyung Jeong;Kim, Soo Hyung;Lee, Guee Sang;Na, In Seop;Kim, Sun Hee
    • International Journal of Contents
    • /
    • v.10 no.1
    • /
    • pp.1-11
    • /
    • 2014
  • In the current generation of smart mobile devices, object tracking is one of the most important research topics for computer vision. Because human face tracking can be widely used for many applications, collecting a dataset of face videos is necessary for evaluating the performance of a tracker and for comparing different approaches. Unfortunately, the well-known benchmark datasets of face videos are not sufficiently diverse. As a result, it is difficult to compare the accuracy between different tracking algorithms in various conditions, namely illumination, background complexity, and subject movement. In this paper, we propose a new dataset that includes 91 face video clips that were recorded in different conditions. We also provide a semi-automatic ground-truth generation tool that can easily be used to evaluate the performance of face tracking systems. This tool helps to maintain the consistency of the definitions for the ground-truth in each frame. The resulting video data set is used to evaluate well-known approaches and test their efficiency.