• Title/Summary/Keyword: Large Scale Data

Search Result 2,796, Processing Time 0.042 seconds

Risk-Incorporated Trajectory Prediction to Prevent Contact Collisions on Construction Sites

  • Rashid, Khandakar M.;Datta, Songjukta;Behzadan, Amir H.;Hasan, Raiful
    • Journal of Construction Engineering and Project Management
    • /
    • v.8 no.1
    • /
    • pp.10-21
    • /
    • 2018
  • Many construction projects involve a plethora of safety-related problems that can cause loss of productivity, diminished revenue, time overruns, and legal challenges. Incorporating data collection and analytics methods can help overcome the root causes of many such problems. However, in a dynamic construction workplace collecting data from a large number of resources is not a trivial task and can be costly, while many contractors lack the motivation to incorporate technology in their activities. In this research, an Android-based mobile application, Preemptive Construction Site Safety (PCS2) is developed and tested for real-time location tracking, trajectory prediction, and prevention of potential collisions between workers and site hazards. PCS2 uses ubiquitous mobile technology (smartphones) for positional data collection, and a robust trajectory prediction technique that couples hidden Markov model (HMM) with risk-taking behavior modeling. The effectiveness of PCS2 is evaluated in field experiments where impending collisions are predicted and safety alerts are generated with enough lead time for the user. With further improvement in interface design and underlying mathematical models, PCS2 will have practical benefits in large scale multi-agent construction worksites by significantly reducing the likelihood of proximity-related accidents between workers and equipment.

Development of a Knowledge Base for Korean Pharmacogenomics Research Network

  • Park, Chan Hee;Lee, Su Yeon;Jung, Yong;Park, Yu Rang;Lee, Hye Won;Kim, Ju Han
    • Genomics & Informatics
    • /
    • v.3 no.3
    • /
    • pp.68-73
    • /
    • 2005
  • Pharmacogenomics research requires an intelligent integration of large-scale genomic and clinical data with public and private knowledge resources. We developed a web-based knowledge base for KPRN (Korea Pharmacogenomics Research Network, http://kprn.snubi. org/). Four major types of information is integrated; genetic variation, drug information, disease information, and literature annotation. Eighteen Korean pharmacogenomics research groups in collaboration have submitted 859 genotype data sets for 91 disease-related genes. Integrative analysis and visualization of the large collection of data supported by integrated biomedical path­ways and ontology resources are provided with a user-friendly interface and visualization engine empowered by Generic Genome Browser.

Improvement of RocksDB Performance via Large-Scale Parameter Analysis and Optimization

  • Jin, Huijun;Choi, Won Gi;Choi, Jonghwan;Sung, Hanseung;Park, Sanghyun
    • Journal of Information Processing Systems
    • /
    • v.18 no.3
    • /
    • pp.374-388
    • /
    • 2022
  • Database systems usually have many parameters that must be configured by database administrators and users. RocksDB achieves fast data writing performance using a log-structured merged tree. This database has many parameters associated with write and space amplifications. Write amplification degrades the database performance, and space amplification leads to an increased storage space owing to the storage of unwanted data. Previously, it was proven that significant performance improvements can be achieved by tuning the database parameters. However, tuning the multiple parameters of a database is a laborious task owing to the large number of potential configuration combinations. To address this problem, we selected the important parameters that affect the performance of RocksDB using random forest. We then analyzed the effects of the selected parameters on write and space amplifications using analysis of variance. We used a genetic algorithm to obtain optimized values of the major parameters. The experimental results indicate an insignificant reduction (-5.64%) in the execution time when using these optimized values; however, write amplification, space amplification, and data processing rates improved considerably by 20.65%, 54.50%, and 89.68%, respectively, as compared to the performance when using the default settings.

Big Data Analysis Using on Based Social Network Service Data (소셜네트워크서비스 기반 데이터를 이용한 빅데이터 분석)

  • Nam, Soo-Tai;Shin, Seong-Yoon;Jin, Chan-Yong
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2019.05a
    • /
    • pp.165-166
    • /
    • 2019
  • Big data analysis is the ability to collect, store, manage and analyze data from existing database management tools. Big data refers to large scale data that is generated in a digital environment, is large in size, has a short generation cycle, and includes not only numeric data but also text and image data. Big data is data that is difficult to manage and analyze in the conventional way. It has huge size, various types, fast generation and velocity. Therefore, companies in most industries are making efforts to create value through the application of Big data. In this study, we analyzed the meaning of keyword using Social Matrix, a big data analysis tool of Daum communications. Also, the theoretical implications are presented based on the analysis results.

  • PDF

A development of hierarchical bayesian model for changing point analysis at watershed scale (유역단위에서의 연강수량의 변동점 분석을 위한 계층적 Bayesian 분석기법 개발)

  • Kim, Jin-Guk;Kim, Jin-Young;Kim, Yoon-Hee;Kwon, Hyun-Han
    • Journal of Korea Water Resources Association
    • /
    • v.50 no.2
    • /
    • pp.75-87
    • /
    • 2017
  • In recent decades, extreme events have been significantly increased over the Korean Peninsula due to climate variability and climate change. The potential changes in hydrologic cycle associated with the extreme events increase uncertainty in water resources planning and designing. For these reasons, a reliable changing point analysis is generally required to better understand regime changes in hydrologic time series at watershed scale. In this study, a hierarchical changing point analysis approach that can apply in a watershed scale is developed by combining the existing changing point analysis method and hierarchical Bayesian method. The proposed model was applied to the selected stations that have annual rainfall data longer than 40 years. The results showed that the proposed model can quantitatively detect the shift in precipitation in the middle of 1990s and identify the increase in annual precipitation compared to the several decades prior to the 1990s. Finally, we explored the changes in precipitation and sea level pressure in the context of large-scale climate anomalies using reanalysis data, for a given change point. It was concluded that the identified large-scale patterns were substantially different from each other.

CNN based data anomaly detection using multi-channel imagery for structural health monitoring

  • Shajihan, Shaik Althaf V.;Wang, Shuo;Zhai, Guanghao;Spencer, Billie F. Jr.
    • Smart Structures and Systems
    • /
    • v.29 no.1
    • /
    • pp.181-193
    • /
    • 2022
  • Data-driven structural health monitoring (SHM) of civil infrastructure can be used to continuously assess the state of a structure, allowing preemptive safety measures to be carried out. Long-term monitoring of large-scale civil infrastructure often involves data-collection using a network of numerous sensors of various types. Malfunctioning sensors in the network are common, which can disrupt the condition assessment and even lead to false-negative indications of damage. The overwhelming size of the data collected renders manual approaches to ensure data quality intractable. The task of detecting and classifying an anomaly in the raw data is non-trivial. We propose an approach to automate this task, improving upon the previously developed technique of image-based pre-processing on one-dimensional (1D) data by enriching the features of the neural network input data with multiple channels. In particular, feature engineering is employed to convert the measured time histories into a 3-channel image comprised of (i) the time history, (ii) the spectrogram, and (iii) the probability density function representation of the signal. To demonstrate this approach, a CNN model is designed and trained on a dataset consisting of acceleration records of sensors installed on a long-span bridge, with the goal of fault detection and classification. The effect of imbalance in anomaly patterns observed is studied to better account for unseen test cases. The proposed framework achieves high overall accuracy and recall even when tested on an unseen dataset that is much larger than the samples used for training, offering a viable solution for implementation on full-scale structures where limited labeled-training data is available.

Survey for Reclaimed Lands in Western Coast of North Korea using Satellite Image data (인공위성 영상 자료를 이용한 북한 지역의 간척지 조사)

  • 신석효;김상철;안기원;김남식
    • Proceedings of the Korean Society of Surveying, Geodesy, Photogrammetry, and Cartography Conference
    • /
    • 2004.04a
    • /
    • pp.251-257
    • /
    • 2004
  • The Electro-Optical Camera(EOC) image of the first Korea Multi-Purpose Satellitel(KOMPSAT-1) has both high resolution and convenient acquisition of research data, but on the other hand it has a defect of one band image. Fortunately, the multispectral Moderate Resolution Imaging Spectroradiometer (MODIS) image data are receiving every day at the Korea Aerospace Research Institute (KARI). Therefore, this paper performed an effective merging for survey of reclaimed land using the high-resolution (6.6m) KOMPSAT-1 EOC image and the multispectral MODIS image data. According this paper prepared map of reclaimed lands in Western Coast of North Korea as quantitative(position and form) survey of reclaimed lands of North Korea using merged image. The use of KOPSAT-1 EOC image and MODIS images was found to be economical such using of large scale areas as reclaimed land or according easy to collect information and such north korea as inaccessible areas like as receiving every day.

  • PDF

Weather Radar Image Gener ation Method Using Inter polation based on CUDA

  • Yang, Liu;Jang, Bong-Joo;Lim, Sanghun;Kwon, Ki-Chang;Lee, Suk-Hwan;Kwon, Ki-Ryong
    • Journal of Korea Multimedia Society
    • /
    • v.18 no.4
    • /
    • pp.473-482
    • /
    • 2015
  • Doppler weather radar is an important tool for meteorological research. Through several decades of development, Doppler weather radar has enormous progress in understanding, detection and warning of meso and micro scale weather system. It makes a significant contribution to weather forecast and weather disaster warning. But the large amount of data process limits the application of Doppler weather radar. This paper proposed for fast weather radar data processing based on CUDA. CDUA is a powerful platform for highly parallel programming developed by NVIDIA. Through running plenty of threads, radar data can be calculated at same time. In experiment, CUDA parallel program can significantly improve weather data processing time.

The Design and Implementation of Electronic Catalog System based on XML for 3-D CAD (3차원 CAD를 위한 XML 기반 전자 카타로그 시스템 설계 및 구현)

  • 이권일;권영희
    • Proceedings of the IEEK Conference
    • /
    • 2003.07d
    • /
    • pp.1609-1612
    • /
    • 2003
  • We designed and implemented electronic catalog system for 3D CAD using XML. The XML(extensible Markup Language) is a Markup Language to describe data structure. XML was originally designed to meet the challenges of large-scale electronic publishing. XML Is also playing an increasingly important role in the exchange of a wide variety of data on the Web and elsewhere. This XML catalog helps you easily create and edit the construction materials data. If you want to obtain informations from the construction materials data which is used by object-oriented 3-D CAD program then this system helps you.

  • PDF

Design and Implementation of A Video Information Management System for Digital Libraries (디지털 도서관을 위한 동영상 정보 관리 시스템의 설계 및 구현)

  • 김현주;권재길;정재희;김인홍;강현석;배종민
    • Journal of Korea Multimedia Society
    • /
    • v.1 no.2
    • /
    • pp.131-141
    • /
    • 1998
  • Video data occurred in multimedia documents consist of a large scale of irregular data including audio-visual, spatial-temporal, and semantic information. In general, it is difficult to grasp the exact meaning of such a video information because video data apparently consist of unmeaningful symbols and numerics. In order to relieve these difficulties, it is necessary to develop an integrated manager for complex structures of video data and provide users of video digital libraries with easy, systematic access mechanisms to video informations. This paper proposes a generic integrated video information model(GIVIM) based on an extended Dublin Core metadata system to effectively store and retrieve video documents in digital libraries. The GIVIM is an integrated mo이 of a video metadata model(VMN) and a video architecture information model(VAIM). We also present design and implementation results of a video document management system(VDMS) based on the GIVIM.

  • PDF