• Title/Summary/Keyword: Generate Data

Search Result 3,066, Processing Time 0.028 seconds

CUTIG: An Automated C Unit Test Data Generator Using Static Analysis (CUTIG: 정적 분석을 이용한 C언어 단위 테스트 데이타 추출 자동화 도구)

  • Kim, Taek-Su;Park, Bok-Nam;Lee, Chun-Woo;Kim, Ki-Moon;Seo, Yun-Ju;Wu, Chi-Su
    • Journal of KIISE:Software and Applications
    • /
    • v.36 no.1
    • /
    • pp.10-20
    • /
    • 2009
  • As unit testing should be performed repeatedly and continuously, it is a high-cost software development activity. Although there are many studies on unit test automation, there are less studies on automated test case generation which are worthy of note. In this paper, we discuss a study on automated test data generation from source codes and indicate algorithms for each stage. We also show some issues of test data generation and introduce an automated test data generating tool: CUTIG. As CUTIG generates test data not from require specifications but from source codes, software developers could generate test data when specifications are insufficient or discord with real implementation. Moreover we hope that the tool could help software developers to reduce cost for test data preparation.

Development of 3D Modeling Technology of Human Vacancy for Bio-CAD (Bio-CAD를 위한 인체공동부의 3차원 모델링 기술 개발)

  • Kim, Ho-Chan;Bae, Yong-Hwan;Kwon, Ki-Su;Seo, Tae-Won;Lee, Seok-Hee
    • Journal of the Korean Society for Precision Engineering
    • /
    • v.26 no.12
    • /
    • pp.138-145
    • /
    • 2009
  • Custom medical treatment is being widely adapted to lots of medical applications. A technology for 3D modeling is strongly required to fabricate medical implants for individual patient. Needs on true 3D CAD data of a patient is strongly required for tissue engineering and human body simulations. Medical imaging devices show human inner section and 3D volume rendering images of human organs. CT or MRI is one of the popular imaging devices for that use. However, those image data is not sufficient to use for medical fabrication or simulation. This paper mainly deals how to generate 3D geometry data from those medical images. A new image processing technology is introduced to reconstruct 3D geometry of a human body vacancy from the medical images. Then a surface geometry data is reconstructed by using Marching cube algorithm. Resulting CAD data is a custom 3D geometry data of human vacancy. This paper introduces a novel 3D reconstruction process and shows some typical examples with implemented software.

Combining Support Vector Machine Recursive Feature Elimination and Intensity-dependent Normalization for Gene Selection in RNAseq (RNAseq 빅데이터에서 유전자 선택을 위한 밀집도-의존 정규화 기반의 서포트-벡터 머신 병합법)

  • Kim, Chayoung
    • Journal of Internet Computing and Services
    • /
    • v.18 no.5
    • /
    • pp.47-53
    • /
    • 2017
  • In past few years, high-throughput sequencing, big-data generation, cloud computing, and computational biology are revolutionary. RNA sequencing is emerging as an attractive alternative to DNA microarrays. And the methods for constructing Gene Regulatory Network (GRN) from RNA-Seq are extremely lacking and urgently required. Because GRN has obtained substantial observation from genomics and bioinformatics, an elementary requirement of the GRN has been to maximize distinguishable genes. Despite of RNA sequencing techniques to generate a big amount of data, there are few computational methods to exploit the huge amount of the big data. Therefore, we have suggested a novel gene selection algorithm combining Support Vector Machines and Intensity-dependent normalization, which uses log differential expression ratio in RNAseq. It is an extended variation of support vector machine recursive feature elimination (SVM-RFE) algorithm. This algorithm accomplishes minimum relevancy with subsets of Big-Data, such as NCBI-GEO. The proposed algorithm was compared to the existing one which uses gene expression profiling DNA microarrays. It finds that the proposed algorithm have provided as convenient and quick method than previous because it uses all functions in R package and have more improvement with regard to the classification accuracy based on gene ontology and time consuming in terms of Big-Data. The comparison was performed based on the number of genes selected in RNAseq Big-Data.

Data BILuring Method for Solving Sparseness Problem in Collaborative Filtering (협동적 여과에서의 희소성 문제 해결을 위한 데이타 블러링 기법)

  • Kim, Hyung-Il;Kim, Jun-Tae
    • Journal of KIISE:Software and Applications
    • /
    • v.32 no.6
    • /
    • pp.542-553
    • /
    • 2005
  • Recommendation systems analyze user preferences and recommend items to a user by predicting the user's preference for those items. Among various kinds of recommendation methods, collaborative filtering(CF) has been widely used and successfully applied to practical applications. However, collaborative filtering has two inherent problems: data sparseness and the cold-start problems. If there are few known preferences for a user, it is difficult to find many similar users, and therefore the performance of recommendation is degraded. This problem is more serious when a new user is first using the system. In this paper we propose a method of integrating additional feature information of users and items into CF to overcome the difficulties caused by sparseness and improve the accuracy of recommendation. In our method, we first fill in unknown preference values by using the probability distribution of feature values, then generate the top-N recommendations by applying collaborative filtering on the modified data. We call this method of filling unknown preference values as data blurring. Several experimental results that show the effectiveness of the proposed method are also presented.

A Study On The Design of Patient Monitoring System Using RFID/WSN Based on Complex Event Processing (복합 이벤트 처리기반 RFID/WSN을 이용한 환자모니터링 시스템 설계에 관한 연구)

  • Park, Yong-Min;Oh, Young-Hwan
    • Journal of the Institute of Electronics Engineers of Korea TC
    • /
    • v.46 no.10
    • /
    • pp.1-7
    • /
    • 2009
  • Nowadays there are many studies and there's huge development about RFID and WSN which have great developmental potential to many kinds of applications. In particular, the healthcare field is expected to could be securing international competitive power in u-Healthcare and combined medical treatment industry and service. More and more real time application apply RFID and WSN technology to identify, data collect and locate objects. Wide deployment of RFID and WSN will generate an unprecedented volume of primitive data in a short time. Duplication and redundancy of primitive data will affect real time performance of application. Thus, emerging applications must filter primitive data and correlate them for complex pattern detection and transform them to events that provide meaningful, actionable information to end application. In this paper, we design a complex event processing system. This system will process RFID and WSN primitive data and event and perform data transformation. Integrate RFID and WSN system had applied each now in medical treatment through this study and efficient data transmission and management forecast that is possible.

Vector Data Hashing Using Line Curve Curvature (라인 곡선 곡률 기반의 벡터 데이터 해싱)

  • Lee, Suk-Hwan;Kwon, Ki-Ryong
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.36 no.2C
    • /
    • pp.65-77
    • /
    • 2011
  • With the rapid expansion of application fields of vector data model such as CAD design drawing and GIS digital map, the security technique for vector data model has been issued. This paper presents the vector data hashing for the authentication and copy protection of vector data model. The proposed hashing groups polylines in main layers of a vector data model and generates the group coefficients by the line curve curvatures of the first and second type of all poly lines. Then we calculate the feature coefficients by projecting the group coefficients onto the random pattern and generate finally the binary hash from the binarization of the feature coefficients. From experimental results using a number of CAD drawings and GIS digital maps, we verified that the proposed hashing has the robustness against various attacks and the uniqueness and security by the random key.

An improvement plan of information system operational audit for database operational management based on data quality (데이터 품질에 기반을 둔 데이터베이스 운영관리를 위한 정보시스템 운영감리 개선 방안)

  • Jang, WonJae;Kim, Dongsoo;Min, Dukki
    • Journal of Service Research and Studies
    • /
    • v.8 no.2
    • /
    • pp.41-65
    • /
    • 2018
  • With the dawn of society where individuals or enterprises based on data generate infinite profits, the significance of database operation management is growing centering on data quality. However, there are not many South Korean public or private entities managing them systematically. Against this backdrop, this study sought to investigate the current status and problems and explore how to improve from the perspective of auditors. To implement this study, audit checklist was improved and, based on it, auditors and IT experts were surveyed. The final data were analyzed to test the study hypotheses empirically. As a result of the analysis, it was found that the auditors had been highly satisfied with all of the items on the improved audit checklist for data quality-based database operation management. Moreover, non-auditors were also found to regard them within their acceptable range. This study is expected to help improve information system operation audit and enterprises data operation management.

Rapid Estimation of the Aerodynamic Coefficients of a Missile via Co-Kriging (코크리깅을 활용한 신속한 유도무기 공력계수 추정)

  • Kang, Shinseong;Lee, Kyunghoon
    • Journal of the Korean Society for Aeronautical & Space Sciences
    • /
    • v.48 no.1
    • /
    • pp.13-21
    • /
    • 2020
  • Surrogate models have been used for the rapid estimation of six-DOF aerodynamic coefficients in the context of the design and control of a missile. For this end, we may generate highly accurate surrogate models with a multitude of aerodynamic data obtained from wind tunnel tests (WTTs); however, this approach is time-consuming and expensive. Thus, we aim to swiftly predict aerodynamic coefficients via co-Kriging using a few WTT data along with plenty of computational fluid dynamics (CFD) data. To demonstrate the excellence of co-Kriging models based on both WTT and CFD data, we first generated two surrogate models: co-Kriging models with CFD data and Kriging models without the CFD data. Afterwards, we carried out numerical validation and examined predictive trends to compare the two different surrogate models. As a result, we found that the co-Kriging models produced more accurate aerodynamic coefficients than the Kriging models thanks to the assistance of CFD data.

Implementation of an Automatic Test Data Generating Tool for Digital TV Software (디지털 TV 소프트웨어를 위한 테스트 데이터 자동 생성기의 구현)

  • Gwak, Tae-Hee;Choi, Byoung-Ju
    • Journal of KIISE:Computing Practices and Letters
    • /
    • v.8 no.5
    • /
    • pp.551-562
    • /
    • 2002
  • Digital TV software, receiver system for digital broadcasting, processes huge MPEG-2 TS formatted data that has variable hierarchy. Because of complexity and enormity of MPEG-2 TS, it is difficult for user to generate test data manually. Generating of test data is not only expensive and time consuming but also requires expert knowledge of MPEG-2 standard. In this paper, we implemented the tool that generates the MPEG-2 TS formatted test data for Digital TV software. Using this tool, user ran get reliable test data without extensive knowledge of MPEG-2 standard. Also, database mechanism that our tool based on supports variable hierarchy of MPEG-2 TS.

Creation and clustering of proximity data for text data analysis (텍스트 데이터 분석을 위한 근접성 데이터의 생성과 군집화)

  • Jung, Min-Ji;Shin, Sang Min;Choi, Yong-Seok
    • The Korean Journal of Applied Statistics
    • /
    • v.32 no.3
    • /
    • pp.451-462
    • /
    • 2019
  • Document-term frequency matrix is a type of data used in text mining. This matrix is often based on various documents provided by the objects to be analyzed. When analyzing objects using this matrix, researchers generally select only terms that are common in documents belonging to one object as keywords. Keywords are used to analyze the object. However, this method misses the unique information of the individual document as well as causes a problem of removing potential keywords that occur frequently in a specific document. In this study, we define data that can overcome this problem as proximity data. We introduce twelve methods that generate proximity data and cluster the objects through two clustering methods of multidimensional scaling and k-means cluster analysis. Finally, we choose the best method to be optimized for clustering the object.