• Title/Summary/Keyword: 클러스터 컴퓨터

Search Result 499, Processing Time 0.026 seconds

The Parallelization Effectiveness Analysis of K-DRUM Model (분포형 강우유출모형(K-DRUM)의 병렬화 효과 분석)

  • Chung, Sung-Young;Park, Jin-Hyeog;Hur, Young-Teck;Jung, Kwan-Sue
    • Journal of Korean Society for Geospatial Information Science
    • /
    • v.18 no.4
    • /
    • pp.21-30
    • /
    • 2010
  • In this paper, the parallel distributed rainfall runoff model(K-DRUM) using MPI(Message Passing Interface) technique was developed to solve the problem of calculation time as it is one of the demerits of the distributed model for performing physical and complicated numerical calculations for large scale watersheds. The K-DRUM model which is based on GIS can simulate temporal and spatial distribution of surface flow and sub-surface flow during flood period, and input parameters of ASCII format as pre-process can be extracted using ArcView. The comparison studies were performed with various domain divisions in Namgang Dam watershed in case of typoon 'Ewiniar' at 2006. The numerical simulation using the cluster system was performed to check a parallelization effectiveness increasing the domain divisions from 1 to 25. As a result, the computer memory size reduced and the calculation time was decreased with increase of divided domains. And also, the tool was suggested in order to decreasing the discharge error on each domain connections. The result shows that the calculation and communication times in each domain have to repeats three times at each time steps in order to minimization of discharge error.

Recognition of Passport Image Using Removing Noise Branches and Enhanced Fuzzy ART (잡영 가지 제거 알고리즘과 개선된 퍼지 ART를 이용한 여권 코드 인식)

  • Lee, Sang-Soo;Jang, Do-Won;Kim, Kwang-Baek
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • v.9 no.2
    • /
    • pp.377-382
    • /
    • 2005
  • 본 논문에서는 출입국자 관리의 효율성과 체계적인 출입국 관리를 위하여 여권 코드를 자동으로 인식하는 방법을 제안한다. 여권 이미지는 기울어진 상태로 스캔 되어 획득되어질 수도 있으므로 기울기 보정은 문자 분할 및 인식에 있어 매우 중요하다. 따라서 본 논문에서는 여권 영상을 스미어링한 후, 추출된 문자열 중에서 가장 긴 문자열을 선택하고 이 문자열의 좌측과 우측 부분의 두께 중심을 연결하는 직선과 수평선과의 기울기를 이용하여 여권 영상에 대한 각도 보정을 수행한다. 여권 코드 추출은 소벨 연산자와 수평 스미어링, 8방향 윤관선 추적 알고리즘을 적용하여 여권 코드의 문자열 영역을 추출하고, 추출된 여권 코드 문자열 영역에 대해 반복 이진화 방법을 적용하여 코드의 문자열 영역을 이진화 한다, 이진화된 문자열 영역에 대해 여권 코드의 인식율을 높이기 위하여 잡영 가지 제거 알고리즘을 적용하여 개별 문자의 잡영을 제거한 후에 개별 코드를 추출하며, CDM 마스크를 적용하여 추출된 개별코드를 복원한다. 추출된 개별코드는 개선된 퍼지 ART 알고리즘을 제안하여 인식에 적용한다. 실제 여권 영상을 대상으로 실험한 결과, CDM 마스크를 이용하여 추출된 개별 코드를 개선된 퍼지 ART 알고리즘을 적용하여 인식한 방법보다 잡영 제거 알고리즘과 CDM 마스크를 적용하여 개선된 퍼지 ART 알고리즘으로 개별 코드를 인식하는 것이 효율적인 것을 확인하였다. 그리고 기존의 퍼지 ART 알고리즘을 이용하여 개별 코드를 인식하는 경우보다 본 논문에서 제안한 개선된 퍼지 ART 알고리즘을 이용하여 개별 코드를 인식하는 경우가 서로 다른 패턴들이 같은 클러스터로 분류되지 않아 인식 성능이 개선되었다.생산하고 있다. 또한 이러한 자료를 바탕으로 지역통계 수요에 즉각 대처할 수 있다. 더 나아가 이와 같은 통계는 전 국민에 대한 패널자료이기 때문에 통계적 활용의 범위가 방대하다. 특히 개인, 가구, 사업체 등 사회 활동의 주체들이 어떻게 변화하는지를 추적할 수 있는 자료를 생산함으로써 다양한 인과적 통계분석을 할 수 있다. 행정자료를 활용한 인구센서스의 이러한 특징은 국가의 교육정책, 노동정책, 복지정책 등 다양한 정책을 정확한 자료를 근거로 수립할 수 있는 기반을 제공한다(Gaasemyr, 1999). 이와 더불어 행정자료 기반의 인구센서스는 비용이 적게 드는 장점이 있다. 예를 들어 덴마크나 핀란드에서는 조사로 자료를 생산하던 때의 1/20 정도 비용으로 행정자료로 인구센서스의 모든 자료를 생산하고 있다. 특히, 최근 모든 행정자료들이 정보통신기술에 의해 데이터베이스 형태로 바뀌고, 인터넷을 근간으로 한 컴퓨터네트워크가 발달함에 따라 각 부처별로 행정을 위해 축적한 자료를 정보통신기술로 연계${cdot}$통합하면 막대한 조사비용을 들이지 않더라도 인구센서스자료를 적은 비용으로 생산할 수 있는 근간이 마련되었다. 이렇듯 행정자료 기반의 인구센서스가 많은 장점을 가졌지만, 그렇다고 모든 국가가 당장 행정자료로 인구센서스를 대체할 수 있는 것은 아니다. 행정자료로 인구센서스통계를 생산하기 위해서는 각 행정부서별로 사용하는 행정자료들을 연계${cdot}$통합할 수 있도록 국가사회전반에 걸쳐 행정 체제가 갖추어져야 하기 때문이다. 특히 모든 국민 개개인에 관한 기본정보, 개인들이 거주하며 생활하는 단위인 개별 주거단위에 관한 정보가 행정부에 등록되어

  • PDF

A RSU-Aided Resource Search and Cloud Construction Mechanism in VANETs (차량 네트워크에서 RSU를 이용한 리소스 검색 및 클라우드 구축 방안)

  • Lee, Yoonhyeong;Lee, Euisin
    • KIPS Transactions on Computer and Communication Systems
    • /
    • v.9 no.3
    • /
    • pp.67-76
    • /
    • 2020
  • With the fast development in wireless communications and vehicular technologies, vehicular ad hoc networks (VANETs) have enabled to deliver data between vehicles. Recently, VANETs introduce a Vehicular Cloud (VC) model for collaborating to share and use resources of vehicles to create value-added services. To construct a VC, a vehicle should search vehicles that intend to provide their own resource. The single-hop search cannot search enough provider vehicles due to a small coverage and non-line-of-sights of communications. On the other hand, the multi-hop search causes very high traffics for large coverage searching and frequent connection breakages. Recently, many Roadside Units (RSUs) have been deployed on roads to collect the information of vehicles in their own coverages and to connect them to Internet. Thus, we propose a RSU-aided vehicular resource search and cloud construction mechanism in VANETS. In the proposed mechanism, a RSU collects the information of location and mobility of vehicles and selects provider vehicles enabled to provide resources needed for constructing a VC of a requester vehicle based on the collected information. In the proposed mechanism, the criteria for determining provider vehicles to provide resources are the connection duration between each candidate vehicle and the requester vehicle, the resource size of each candidate vehicle, and its connection starting time to the requester vehicle. Simulation results verify that the proposed mechanism achieves better performance than the existing mechanism.

Development of a Scalable Clustering A/V Server for the Internet Personal-Live Broadcasting (인터넷 개인 생방송을 위한 Scalable Clustering A/V Server 개발)

  • Lee, Sang-Moon;Kang, Sin-Jun;Min, Byung-Seok;Kim, Hag-Bae;Park, Jin-Bae
    • The KIPS Transactions:PartC
    • /
    • v.9C no.1
    • /
    • pp.107-114
    • /
    • 2002
  • In these days, rapid advances of the computer system and the high speed network have made the multimedia services popularized among various applications and services in the internet. Internet live broadcasting, a part of multimedia services, makes it possible to provide not only existing broadcasting services including audio and video but also interactive communications which also expand application scopes by freeing from both temporal and spatial limitation. In the Paper, an interned Personal-live broadcasting server system is developed by allowing individual users to actively create or join live-broadcasting services with such basic multimedia devices as a PC camera and a sound card. As the number of broadcasters and participants increases, concurrent multiple channels are established and groups are to be expanded. The system should also guarantee High Availability (HA) for continuous services even in the presence of partial failure of the cluster. Furthermore, a transmission mode switching is supported to consider network environments in the user system.

A Study on Optimum Coding Method for Correlation Processing of Radio Astronomy (전파천문 상관처리를 위한 최적 코딩 방법에 관한 연구)

  • Shin, Jae-Sik;Oh, Se-Jin;Yeom, Jae-Hwan;Roh, Duk-Gyoo;Chung, Dong-Kyu;Oh, Chung-Sik;Hwang, Ju-Yeon;So, Yo-Hwan
    • Journal of the Institute of Convergence Signal Processing
    • /
    • v.16 no.4
    • /
    • pp.139-148
    • /
    • 2015
  • In this paper, the optimum coding method is proposed by using open library in order to improve the performance of a software correlator developed for Korea-Japan Joint VLBI Correlator(KJJVC). The correlation system for VLBI observing system is generally implemented with hardware using ASIC or FPGA because the computational quantity is increased geometrically according to the participated observatory number. However, the software correlation system is recently constructed at a massive server such as a cluster using software according to the development of computing power. Since VLBI correlator implemented with hardware is able to conduct data processing with real-time or quasi real-time compared with mostly observational time, software correlation has to perform optimal data processing in coding work so as to have the same performance as that of the hardware. Therefore, in this paper, the experimental comparison was conducted by open-source based fftw library released in FFT processing stage, which is the most important part of the correlator system for performing optimum coding work in software development phase, such as general method using fftw library or methods using SSE(Streaming SIMD Extensions), shared memory, or OpenMP, and method using merged techniques listed above. Through the experimental results, the proposed optimum coding method for improving the performance of developed software correlator using fftw library, shared memory and OpenMP is effectively confirmed by reducing correlation time compared with conventional method.

Design and Implementation of a Benchmarking System Based on ArangoDB (ArangoDB기반 벤치마킹 시스템 설계 및 구현)

  • Choi, Do-Jin;Baek, Yeon-Hee;Lee, So-Min;Kim, Yun-A;Kim, Nam-Young;Choi, Jae-Young;Lee, Hyeon-Byeong;Lim, Jong-Tae;Bok, Kyoung-Soo;Song, Seok-Il;Yoo, Jae-Soo
    • The Journal of the Korea Contents Association
    • /
    • v.21 no.9
    • /
    • pp.198-208
    • /
    • 2021
  • ArangoDB is a NoSQL database system that has been popularly utilized in many applications for storing large amounts of data. In order to apply a new NoSQL database system such as ArangoDB, to real work environments we need a benchmarking system that can evaluate its performance. In this paper, we design and implement a ArangoDB based benchmarking system that measures a kernel level performance well as an application level performance. We partially modify YCSB to measure the performance of a NoSQL database system in the cluster environment. We also define three real-world workload types by analyzing the existing materials. We prove the feasibility of the proposed system through the benchmarking of three workload types. We derive available workloads in ArangoDB and show that performance at the kernel layer as well as the application layer can be visualized through benchmarking of three workload types. It is expected that applicability and risk reviews will be possible through benchmarking of this system in environments that need to transfer data from the existing database engine to ArangoDB.

Variable Selection of Feature Pattern using SVM-based Criterion with Q-Learning in Reinforcement Learning (SVM-기반 제약 조건과 강화학습의 Q-learning을 이용한 변별력이 확실한 특징 패턴 선택)

  • Kim, Chayoung
    • Journal of Internet Computing and Services
    • /
    • v.20 no.4
    • /
    • pp.21-27
    • /
    • 2019
  • Selection of feature pattern gathered from the observation of the RNA sequencing data (RNA-seq) are not all equally informative for identification of differential expressions: some of them may be noisy, correlated or irrelevant because of redundancy in Big-Data sets. Variable selection of feature pattern aims at differential expressed gene set that is significantly relevant for a special task. This issues are complex and important in many domains, for example. In terms of a computational research field of machine learning, selection of feature pattern has been studied such as Random Forest, K-Nearest and Support Vector Machine (SVM). One of most the well-known machine learning algorithms is SVM, which is classical as well as original. The one of a member of SVM-criterion is Support Vector Machine-Recursive Feature Elimination (SVM-RFE), which have been utilized in our research work. We propose a novel algorithm of the SVM-RFE with Q-learning in reinforcement learning for better variable selection of feature pattern. By comparing our proposed algorithm with the well-known SVM-RFE combining Welch' T in published data, our result can show that the criterion from weight vector of SVM-RFE enhanced by Q-learning has been improved by an off-policy by a more exploratory scheme of Q-learning.

Time Synchronization Robust to Topology Change Through Reference Node Re-Election (기준노드의 재선정을 통한 토폴로지 변화에 강인한 시간 동기화)

  • Jeon, Young;Kim, Taehong;Kim, Taejoon;Lee, Jaeseang;Ham, Jae-Hyun
    • KIPS Transactions on Computer and Communication Systems
    • /
    • v.8 no.8
    • /
    • pp.191-200
    • /
    • 2019
  • In an Ad-hoc network, a method of time synchronizing all the nodes in a network centering on one reference node can be used. A representative algorithm based on a reference node is Flooding Time Synchronization Protocol (FTSP). In the process of sending and receiving messages, predictable and unpredictable delays occur, which should be removed because it hinders accurate time synchronization. In multi-hop communications, hop delays occur when a packet traverses a number of hops. These hop delays significantly degrade the synchronization performance among nodes. Therefore, we need to find a method to reduce these hop delays and increase synchronization performance. In the FTSP scheme, hop delays can be greatly increased depending on the position of a reference node. In addition, in FTSP, a node with the smallest node ID is elected as a reference node, hence, the position of a reference node is actually arbitrarily determined. In this paper, we propose an optimal reference node election algorithm to reduce hop delays, and compare the performance of the proposed scheme with FTSP using the network simulator OPNET. In addition, we verify that the proposed scheme has an improved synchronization performance, which is robust to topology changes.

Deep Learning OCR based document processing platform and its application in financial domain (금융 특화 딥러닝 광학문자인식 기반 문서 처리 플랫폼 구축 및 금융권 내 활용)

  • Dongyoung Kim;Doohyung Kim;Myungsung Kwak;Hyunsoo Son;Dongwon Sohn;Mingi Lim;Yeji Shin;Hyeonjung Lee;Chandong Park;Mihyang Kim;Dongwon Choi
    • Journal of Intelligence and Information Systems
    • /
    • v.29 no.1
    • /
    • pp.143-174
    • /
    • 2023
  • With the development of deep learning technologies, Artificial Intelligence powered Optical Character Recognition (AI-OCR) has evolved to read multiple languages from various forms of images accurately. For the financial industry, where a large number of diverse documents are processed through manpower, the potential for using AI-OCR is great. In this study, we present a configuration and a design of an AI-OCR modality for use in the financial industry and discuss the platform construction with application cases. Since the use of financial domain data is prohibited under the Personal Information Protection Act, we developed a deep learning-based data generation approach and used it to train the AI-OCR models. The AI-OCR models are trained for image preprocessing, text recognition, and language processing and are configured as a microservice architected platform to process a broad variety of documents. We have demonstrated the AI-OCR platform by applying it to financial domain tasks of document sorting, document verification, and typing assistance The demonstrations confirm the increasing work efficiency and conveniences.