• Title/Summary/Keyword: Data Source

Search Result 6,579, Processing Time 0.033 seconds

Ubiquitous Data Warehosue: Integrating RFID with Mutidimensional Online Analysis (유비쿼터스 데이터 웨어하우스: RFID와 다차원 온라인 분석의 통합)

  • Cho, Dai-Yon;Lee, Seung-Pyo
    • Journal of Information Technology Services
    • /
    • v.4 no.2
    • /
    • pp.61-69
    • /
    • 2005
  • RFID is used for tracking systems in various business fields these days and these systems brought considerable efficiencies and cost savings to companies. Real-time based information acquired through RFID devices could be a valuable source of information for making decisions if it is combined with decision support tools like OLAP of a data warehouse that has originally been designed for analyzing static and historical data. As an effort of extending the data source of a data warehouse, RFID is combined with a data warehouse in this research. And OLAP is designed to analyze the dynamic real-time based information gathered through RFID devices. The implemented prototype shows that ubiquitous computing technology such as RFID could be a valuable data source for a data warehouse and is very useful for making decisions when it is combined with online analysis. The system architecture of such system is suggested.

Implementation and Performance Analysis of Real-time Multi-source Sensor Data Management System Based on Wireless Sensor Network (무선 센서네트워크 기반 실시간 다중소스 센서데이터 관리시스템 구현 및 성능분석)

  • Kang, Moon-Sik
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.36 no.8B
    • /
    • pp.1003-1011
    • /
    • 2011
  • In this paper, a real-time multi-source sensor data management system based on wireless sensor network is proposed and implemented. The proposed management system is designed to transmit the wireless data to the server in order to monitor and control the multi-source target's status efficiently by analyzing them. The proposed system is implemented to make it possible to control and transmit the wireless sensor data by classifying them, of which data are issued from the clustered sources composed of a number of the remote multiple sensors. In order to evaluate the performance of the proposed system, we measure and analyze both the transmission delay time according to the distance and the data loss rate issued from multiple data sources. From these results, it is verified that the proposed system has a good performance.

Visualization Techniques for Massive Source Code (대용량 소스코드 시각화기법 연구)

  • Seo, Dong-Su
    • The Journal of Korean Association of Computer Education
    • /
    • v.18 no.4
    • /
    • pp.63-70
    • /
    • 2015
  • Program source code is a set of complex syntactic information which are expressed in text forms, and contains complex logical structures. Structural and logical complexity inside source code become barriers in applying visualization techniques shown in traditional big-data approaches when the volume of source code become over ten-thousand lines of code. This paper suggests a procedure for making visualization of structural characteristics in source code. For this purpose, this paper defines internal data structures as well as inter-procedural relationships among functions. The paper also suggests a means of outlining the structural characteristics of source code by visualizing the source codes with network forms The result of the research work can be used as a means of controling and understanding the massive volume of source code.

A study for Verification Procedures on Open-source Software Via Benchmark Testing (벤치마크 테스트를 통한 공개소프트웨어 검증 절차에 관한 연구)

  • Kim, Doo-Yeon;Rhew, Sung-Yul
    • Journal of Information Technology Services
    • /
    • v.5 no.3
    • /
    • pp.99-108
    • /
    • 2006
  • Public institutions are considering adoption of open-source software in the process of information projects. However, there doesn't exist reliable information about an adoption process for open-source software. Performance and stability problems of this software also persist, as a result, current open-source software is not widely used. As a software market and industry grows, Benchmark test has been performed more often than before in order to help customers understand and select the most appropriate product among myriad similar ones. It is certain that more objective and trustful data evidence should be obtained by way of utilizing the procedures and methods of Benchmark Test in decision making process for selecting an open-source software. For this research, Benchmark test was applied as a way of demonstrating performance verification of an open-source software in the public institutions. It is certain that more objective and trustful data evidence should be obtained by way of utilizing the procedures and methods of Benchmark Test in decision making process for selecting an open-source software. It also introduces a case study of a information system, which selected and implemented open-source software, in order to confirm the validity of this research. This research will serve as a guideline to adopt open-source software in governments as well as public institutions.

Designing Cost Effective Open Source System for Bigdata Analysis (빅데이터 분석을 위한 비용효과적 오픈 소스 시스템 설계)

  • Lee, Jong-Hwa;Lee, Hyun-Kyu
    • Knowledge Management Research
    • /
    • v.19 no.1
    • /
    • pp.119-132
    • /
    • 2018
  • Many advanced products and services are emerging in the market thanks to data-based technologies such as Internet (IoT), Big Data, and AI. The construction of a system for data processing under the IoT network environment is not simple in configuration, and has a lot of restrictions due to a high cost for constructing a high performance server environment. Therefore, in this paper, we will design a development environment for large data analysis computing platform using open source with low cost and practicality. Therefore, this study intends to implement a big data processing system using Raspberry Pi, an ultra-small PC environment, and open source API. This big data processing system includes building a portable server system, building a web server for web mining, developing Python IDE classes for crawling, and developing R Libraries for NLP and visualization. Through this research, we will develop a web environment that can control real-time data collection and analysis of web media in a mobile environment and present it as a curriculum for non-IT specialists.

An Extended Relational Data Model for Database Uncertainty Using Data Source Reliability (데이터 제공원의 신뢰도를 고려한 확장 관계형 데이터 모델)

  • 정철용;이석균;서용무
    • The Journal of Information Technology and Database
    • /
    • v.6 no.1
    • /
    • pp.15-25
    • /
    • 1999
  • We propose an extended relational data model which can represent the reliability of data. In this paper, the reliability of data is defined as the reliability of the source, from which the data originated. We represent the reliability of data at the level of attribute values, instead of tuples, then define the selection, product and join operators.

  • PDF

Application of Neural Network to Determine the Source Location in Acoustic Emission

  • Lee, Sang-Eun
    • Journal of the Korean Society for Nondestructive Testing
    • /
    • v.25 no.6
    • /
    • pp.475-482
    • /
    • 2005
  • The iterative calculation by least square method was used to determine the source location of acoustic emission in rock, as so called "traditional method". The results were compared with source coordinates infered from the application of neural network system for new input data, as so called "new method". Input data of the neural network were based on the time differences of longitudinal waves arrived from acoustic emission events at each transducer, the variation of longitudinal velocities at each stress level, and the coordinates of transducer as in the traditional method. The momentum back propagation neural network system adopted to determine source location, which consists of three layers, and has twenty-seven input processing elements. Applicability of the new method were identified, since the results of source location by the application of two methods were similarly concordant.

The Method of Elevation Accuracy In Sound Source Localization System (음원 위치 추정 시스템의 정확도 향상 방법)

  • Kim, Yong-Eun;Chung, Jin-Gyun
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.46 no.2
    • /
    • pp.24-29
    • /
    • 2009
  • Sound source localization system is used in a robot, a video conference and CCTV(Closed-circuit television) systems. In this Sound source localization systems are applied to human and they can receive a number of sound data frames during speaking. In this paper, we propose methods which is reducing angle estimation error by selecting sound data frame which can more precisely compute the angles from inputted sound data frame. After selected data converted to angle, the error of sound source localization recognition system can be reduced by applying to medium filter. By the experiment using proposed system it is shown that the average error of angle estimation in sound source recognition system can be reduced up to 31 %.

Priority Analysis for Software Functions Using Social Network Analysis and DEA(Data Envelopment Analysis) (사회연결망 분석과 자료포락분석 기법을 이용한 소프트웨어 함수 우선순위 분석 연구)

  • Huh, Sang Moo;Kim, Woo Je
    • Journal of Information Technology Services
    • /
    • v.17 no.3
    • /
    • pp.171-189
    • /
    • 2018
  • To remove software defects and improve performance of software, many developers perform code inspections and use static analysis tools. A code inspection is an activity that is performed manually to detect software defects in the developed source. However, there is no clear criterion which source codes are inspected. A static analysis tool can automatically detect software defects by analyzing the source codes without running the source codes. However, it has disadvantage that analyzes only the codes in the functions without analyzing the relations among source functions. The functions in the source codes are interconnected and formed a social network. Functions that occupy critical locations in a network can be important enough to affect the overall quality. Whereas, a static analysis tool merely suggests which functions were called several times. In this study, the core functions will be elicited by using social network analysis and DEA (Data Envelopment Analysis) for CUBRID open database sources. In addition, we will suggest clear criteria for selecting the target sources for code inspection and will suggest ways to find core functions to minimize defects and improve performance.

High Resolution Hydroacoustic Investigation in Shallow Water for the Engineering Design of Railroad Bridge (철도교량 설계 지반조사를 위한 고분해능 수면 탄성파반사법의 응용 사례)

  • ;Swoboda Ulrich
    • Proceedings of the Korean Geotechical Society Conference
    • /
    • 2001.03a
    • /
    • pp.231-238
    • /
    • 2001
  • To investigate the underground structure of shallow water, Han-river near Yangsou-Ri, high resolution hydroacoustic measurements were carried out for the engineering design of railroad bridge. The acoustic source was a Boomer with an energy of 90 to 280J and in a frequency range up to about 16KHz. The reflected signals were received by using both traditional hydrophones(passive element) and a specially devised receiver unit(active element) mainly composed of piezofilms and preamplifier. They are connected to the "SUMMIT" data acquisition system(DMT-GeoTec company), where the sampling interval was set to 1/32㎳. The source position was continuously monitored by a precision DGPS system whose positioning accuracy was on the order of loom. For the quality control purposes, two different source-receiver geometries were taken. That is to say, the measurements were repeated along the profile everytime depending on the different source energy(175J, 280J), the receiving elements(passive, active) and two different source-receiver geometries. It was shown that the data resolution derived from a proper arrangement with the active hydrophone could be greatly enhanced and hence the corresponding profile section caused by the regular data processing system "FOCUS" accounted excellently for the underground formation below the shallow water.w the shallow water.

  • PDF