• Title/Summary/Keyword: Data Collecting

Search Result 2,221, Processing Time 0.03 seconds

Changes in Measuring Methods of Walking Behavior and the Potentials of Mobile Big Data in Recent Walkability Researches (보행행태조사방법론의 변화와 모바일 빅데이터의 가능성 진단 연구 - 보행환경 분석연구 최근 사례를 중심으로 -)

  • Kim, Hyunju;Park, So-Hyun;Lee, Sunjae
    • Journal of the Architectural Institute of Korea Planning & Design
    • /
    • v.35 no.1
    • /
    • pp.19-28
    • /
    • 2019
  • The purpose of this study is to evaluate the walking behavior analysis methodology used in the previous studies, paying attention to the demand for empirical data collecting for urban and neighborhood planning. The preceding researches are divided into (1)Recording, (2) Surveys, (3)Statistical data, (4)Global positioning system (GPS) devices, and (5)Mobile Big Data analysis. Next, we analyze the precedent research and identify the changes of the walkability research. (1)being required empirical data on the actual walking and moving patterns of people, (2)beginning to be measured micro-walking behaviors such as actual route, walking facilities, detour, walking area. In addition, according to the trend of research, it is analyzed that the use of GPS device and the mobile big data are newly emerged. Finally, we analyze pedestrian data based on mobile big data in terms of 'application' and distinguishing it from existing survey methodology. We present the possibility of mobile big data. (1)Improvement of human, temporal and spatial constraints of data collection, (2)Improvement of inaccuracy of collected data, (3)Improvement of subjective intervention in data collection and preprocessing, (4)Expandability of walking environment research.

Imputation of Missing Data Based on Hot Deck Method Using K-nn (K-nn을 이용한 Hot Deck 기반의 결측치 대체)

  • Kwon, Soonchang
    • Journal of Information Technology Services
    • /
    • v.13 no.4
    • /
    • pp.359-375
    • /
    • 2014
  • Researchers cannot avoid missing data in collecting data, because some respondents arbitrarily or non-arbitrarily do not answer questions in studies and experiments. Missing data not only increase and distort standard deviations, but also impair the convenience of estimating parameters and the reliability of research results. Despite widespread use of hot deck, researchers have not been interested in it, since it handles missing data in ambiguous ways. Hot deck can be complemented using K-nn, a method of machine learning, which can organize donor groups closest to properties of missing data. Interested in the role of k-nn, this study was conducted to impute missing data based on the hot deck method using k-nn. After setting up imputation of missing data based on hot deck using k-nn as a study objective, deletion of listwise, mean, mode, linear regression, and svm imputation were compared and verified regarding nominal and ratio data types and then, data closest to original values were obtained reasonably. Simulations using different neighboring numbers and the distance measuring method were carried out and better performance of k-nn was accomplished. In this study, imputation of hot deck was re-discovered which has failed to attract the attention of researchers. As a result, this study shall be able to help select non-parametric methods which are less likely to be affected by the structure of missing data and its causes.

A study on the Traffic Density Collect System using View Synthesis and Data Analysis (영상정합을 이용한 교통밀도 수집방법과 수집 데이터 비교분석)

  • Park, Bumjin;Roh, Chang-gyun
    • The Journal of The Korea Institute of Intelligent Transport Systems
    • /
    • v.17 no.5
    • /
    • pp.77-87
    • /
    • 2018
  • Traffic Density is the most important of the three primary macroscopic traffic stream parameters, because it is most directly related to traffic demand(Traffic Engineering, 2004). It is defined as the number of existing vehicles within a given distance at a certain time. However, due to weather, road conditions, and cost issues, collecting density directly on the field is difficult. This makes studies of density less actively than those of traffic volume or velocity. For these reasons, there is insufficient attempts on divers collecting methods or researches on the accuracy of measured values. In this paper, we used the 'Density Measuring System' based on the synthesise technology of several camera images as a method to measure density. The collected density value by the 'Density Mesuring System' is selected as the true value based on the density define, and this value was compared with the density calculated by the traditional measurement methods. As a result of the comparison, the density value using the fundamental equation method is the closest to the true value as RMSE shows 1.8 to 2.5. In addition, we investigated some issues that can be overlooked easily such as the collecting interval to be considered on collecting density directly by calculating the moment density and the average density. Despite the actual traffic situation of the experiment site is LOS B, it is difficult to judge the real traffic situation because the moment density values per second are observed max 16.0 (veh/km) to min 2.0 (veh/km). However, the average density measured for 15 minutes at 30-second intervals was 8.3-7.9 (veh/km) and it indicates precisely LOS B.

Development of Outage Data Management System to Calculate the Probability for KEPCO Transmission Systems (한전계통의 송전망 고장확률 산정을 위한 상정고장 DB 관리시스텀(ezCas) 개발)

  • Cha S. T.;Jeon D. H.;Kim T. K.;Jeon M. R.;Choo J. B.;Kim J. O.;Lee S .H
    • Proceedings of the KIEE Conference
    • /
    • summer
    • /
    • pp.88-90
    • /
    • 2004
  • Data are a critical utility asset. Collecting correct data on site leads to accurate information. Data, when gathered with foresight & properly formatted, are useful to both existing database and easily transferable to newer, more comprehensive historical outage data. However, when investigating data items options, the task, can be an arduous one, often requiring the efforts of entire committees. This paper firstly discusses the KEPCO's past 10 years of historical outage data which include meterological data, and also by several elements of the National Weather Service, failure rate, outage duration, and probability classification, etc. Then, these collected data are automatically stored in an Outage Data Management System (ODMS), which allows for easy access and display. ODMS has a straight-forward and easy-to-use interface. It lets you to navigate through modules very easily and allows insertion, deletion or editing of data. In particular, this will further provide the KEPCO that not only helps with probabilistic security assessment but also provides a platform for future development of Probability Estimation Program (PEP).

  • PDF

Implementation of Search Engine to Minimize Traffic Using Blockchain-Based Web Usage History Management System

  • Yu, Sunghyun;Yeom, Cheolmin;Won, Yoojae
    • Journal of Information Processing Systems
    • /
    • v.17 no.5
    • /
    • pp.989-1003
    • /
    • 2021
  • With the recent increase in the types of services provided by Internet companies, collection of various types of data has become a necessity. Data collectors corresponding to web services profit by collecting users' data indiscriminately and providing it to the associated services. However, the data provider remains unaware of the manner in which the data are collected and used. Furthermore, the data collector of a web service consumes web resources by generating a large amount of web traffic. This traffic can damage servers by causing service outages. In this study, we propose a website search engine that employs a system that controls user information using blockchains and builds its database based on the recorded information. The system is divided into three parts: a collection section that uses proxy, a management section that uses blockchains, and a search engine that uses a built-in database. This structure allows data sovereigns to manage their data more transparently. Search engines that use blockchains do not use internet bots, and instead use the data generated by user behavior. This avoids generation of traffic from internet bots and can, thereby, contribute to creating a better web ecosystem.

Design and Evaluation of a Quorum-Based Adaptive Dissemination Algorithm for Critical Data in IoTs (IoT에서 중요한 데이터를 위한 쿼럼 기반 적응적 전파 알고리즘의 설계 및 평가)

  • Bae, Ihn Han;Noh, Heung Tae
    • Journal of Korea Multimedia Society
    • /
    • v.22 no.8
    • /
    • pp.913-922
    • /
    • 2019
  • The Internet of Things (IoT) envisions smart objects collecting and sharing data at a massive scale via the Internet. One challenging issue is how to disseminate data to relevant data consuming objects efficiently. In such a massive IoT network, Mission critical data dissemination imposes constraints on the message transfer delay between objects. Due to the low power and communication range of IoT objects, data is relayed over multi-hops before arriving at the destination. In this paper, we propose a quorum-based adaptive dissemination algorithm (QADA) for the critical data in the monitoring-based applications of massive IoTs. To design QADA, we first design a new stepped-triangular grid structures (sT-grid) that support data dissemination, then construct a triangular grid overlay in the fog layer on the lower IoT layer and propose the data dissemination algorithm of the publish/subscribe model that adaptively uses triangle grid (T-grid) and sT-grid quorums depending on the mission critical in the overlay constructed to disseminate the critical data, and evaluate its performance as an analytical model.

Technical Analysis of an MRV System in Relation to the Implementation of a Data Collection System by the International Maritime Organization (국제해사기구 데이터수집시스템 도입에 따른 MRV 지원시스템의 기술적 분석)

  • Kang, Nam-seon;Lee, Jung-yup;Hong, Yeon-jeong;Byeon, Sang-su;Kim, Jin-yhyung
    • Journal of the Korean Society of Marine Environment & Safety
    • /
    • v.23 no.1
    • /
    • pp.122-129
    • /
    • 2017
  • This study presents the results from a technical analysis of a portal system that is compatible with MRV regulations and utilized to examine energy efficiency in international shipping, in relation to the implementation of a mandatory data collection system by the International Maritime Organization. The details of the SEEMP guidelines, including the data collection system and methods for collecting data on fuel use, were reviewed. Strategies for domestic shipping companies toward MRV have been recommended by identifying differences with the EU MRV, and the technical adequacy of the MRV system was assessed. The MRV system enhances cost and work efficiency by managing emissions data from the early stage to the final stage. It is capable of collecting and reporting emissions data while adhering to the reporting procedures of shipping companies. By granting different access privileges to users, the system supports shipping companies in their data collection and reporting, and also supports verifiers in their data verification activities. Moreover, it makes possible the submission of reports in electronic from, thereby enabling shipping companies to adopt an integrated response to international MRV regulations.

Development of Data Warehouse Systems to Support Cost Analysis in the Ship Production (조선산업의 비용분석 데이터 웨어하우스 시스템 개발)

  • Hwang, Sung-Ryong;Kim, Jae-Gyun;Jang, Gil-Sang
    • IE interfaces
    • /
    • v.15 no.2
    • /
    • pp.159-171
    • /
    • 2002
  • Data Warehouses integrate data from multiple heterogeneous information sources and transform them into a multidimensional representation for decision support applications. Data warehousing has emerged as one of the most powerful tools in delivering information to users. Most previous researches have focused on marketing, customer service, financing, and insurance industry. Further, relatively less research has been done on data warehouse systems in the complex manufacturing industry such as ship production, which is characterized complex product structures and production processes. In the ship production, data warehouse systems is a requisite for effective cost analysis because collecting and analysis of diverse and large of cost-related(material/production cost, productivity) data in its operational systems, was becoming increasingly cumbersome and time consuming. This paper proposes architecture of the data warehouse systems to support cost analysis in the ship production. Also, in order to illustrate the usefulness of the proposed architecture, the prototype system is designed and implemented with the object of the enterprise of producing a large-scale ship.

PAPG: Private Aggregation Scheme based on Privacy-preserving Gene in Wireless Sensor Networks

  • Zeng, Weini;Chen, Peng;Chen, Hairong;He, Shiming
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.10 no.9
    • /
    • pp.4442-4466
    • /
    • 2016
  • This paper proposes a privacy-preserving aggregation scheme based on the designed P-Gene (PAPG) for sensor networks. The P-Gene is constructed using the designed erasable data-hiding technique. In this P-Gene, each sensory data item may be hidden by the collecting sensor node, thereby protecting the privacy of this data item. Thereafter, the hidden data can be directly reported to the cluster head that aggregates the data. The aggregation result can then be recovered from the hidden data in the cluster head. The designed P-Genes can protect the privacy of each data item without additional data exchange or encryption. Given the flexible generation of the P-Genes, the proposed PAPG scheme adapts to dynamically changing reporting nodes. Apart from its favorable resistance to data loss, the extensive analyses and simulations demonstrate how the PAPG scheme efficiently preserves privacy while consuming less communication and computational overheads.

Dialogic Male Voice Triphone DB Construction (남성 음성 triphone DB 구축에 관한 연구)

  • Kim, Yu-Jin;Baek, Sang-Hoon;Han, Min-Soo;Chung, Jae-Ho
    • The Journal of the Acoustical Society of Korea
    • /
    • v.15 no.2
    • /
    • pp.61-71
    • /
    • 1996
  • In this paper, dialogic triphone data base construction for triphone synthesis system is discussed. Particularly, in this work, dialogic speech data is collected from the broadcast media, and three different transcription steps are taken. Total 10 hours of speech data are collected. Among them, six hours of speech data are used for the triphone data base construction, and the rest four hours of data are reserved. Dialogic speech data base construction is far different from the reciting speech data base construction. This paper describes various steps that necessary for the dialogic triphone data base construction from collecting speech data to triphone unit labeling.

  • PDF