• Title/Summary/Keyword: Cluster computing environment

Search Result 109, Processing Time 0.022 seconds

Group Key Agreement for Mobile Devices in Different Cells (서로 다른 셀의 모바일 장치간의 그룹키 동의 프로토콜)

  • Kim Jeeyeon;Choi Yeonyi;Kim Seungjoo;Won Dongho
    • Journal of KIISE:Information Networking
    • /
    • v.32 no.6
    • /
    • pp.651-658
    • /
    • 2005
  • Mobile communication has become more pervasive and it is considered as one of main concerns oi conferencing, multi-user games and etc. in mobile environment. These applications need to secure communication in group. Most of the published protocols are based on model which consists of a stationary base station and a cluster of mobile devices. In this paper, we have focused on the extended model of which participants are several base stations and mobile devices in different cells. We present a new group key protocol among mobile devices in different cells and analyze its security And we also look at how password authentication can be used to our group key agreement protocol. The mobile device's computing load may be reduced by using password authentication.

Cost-Effective MapReduce Processing in the Cloud (클라우드 환경에서의 비용 효율적인 맵리듀스 처리)

  • Ryu, Wooseok
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2018.10a
    • /
    • pp.114-115
    • /
    • 2018
  • This paper studies a mechanism for cost-effective analysis of big data in the cloud environment. Recently, as a storage of electronic medical records can be managed outside the hospital, there is a growing demand for cloud-based big data analysis in small-and-medium hospitals. This paper firstly analyze the Amazon Elastic MapReduce which is a popular cloud framework for big data analysis, and proposes a cost model for analyzing big data using Amazon EMR with less cost. Using the proposed model, the user can construct a cost-effective computing cluster, which maximize the effectiveness of the analysis per operational cost.

  • PDF

SPQUSAR : A Large-Scale Qualitative Spatial Reasoner Using Apache Spark (SPQUSAR : Apache Spark를 이용한 대용량의 정성적 공간 추론기)

  • Kim, Jongwhan;Kim, Jonghoon;Kim, Incheol
    • KIISE Transactions on Computing Practices
    • /
    • v.21 no.12
    • /
    • pp.774-779
    • /
    • 2015
  • In this paper, we present the design and implementation of a large-scale qualitative spatial reasoner using Apache Spark, an in-memory high speed cluster computing environment, which is effective for sequencing and iterating component reasoning jobs. The proposed reasoner can not only check the integrity of a large-scale spatial knowledge base representing topological and directional relationships between spatial objects, but also expand the given knowledge base by deriving new facts in highly efficient ways. In general, qualitative reasoning on topological and directional relationships between spatial objects includes a number of composition operations on every possible pair of disjunctive relations. The proposed reasoner enhances computational efficiency by determining the minimal set of disjunctive relations for spatial reasoning and then reducing the size of the composition table to include only that set. Additionally, in order to improve performance, the proposed reasoner is designed to minimize disk I/Os during distributed reasoning jobs, which are performed on a Hadoop cluster system. In experiments with both artificial and real spatial knowledge bases, the proposed Spark-based spatial reasoner showed higher performance than the existing MapReduce-based one.

Enhancement of Authentication Performance based on Multimodal Biometrics for Android Platform (안드로이드 환경의 다중생체인식 기술을 응용한 인증 성능 개선 연구)

  • Choi, Sungpil;Jeong, Kanghun;Moon, Hyeonjoon
    • Journal of Korea Multimedia Society
    • /
    • v.16 no.3
    • /
    • pp.302-308
    • /
    • 2013
  • In this research, we have explored personal authentication system through multimodal biometrics for mobile computing environment. We have selected face and speaker recognition for the implementation of multimodal biometrics system. For face recognition part, we detect the face with Modified Census Transform (MCT). Detected face is pre-processed through eye detection module based on k-means algorithm. Then we recognize the face with Principal Component Analysis (PCA) algorithm. For speaker recognition part, we extract features using the end-point of voice and the Mel Frequency Cepstral Coefficient (MFCC). Then we verify the speaker through Dynamic Time Warping (DTW) algorithm. Our proposed multimodal biometrics system shows improved verification rate through combining two different biometrics described above. We implement our proposed system based on Android environment using Galaxy S hoppin. Proposed system presents reduced false acceptance ratio (FAR) of 1.8% which shows improvement from single biometrics system using the face and the voice (presents 4.6% and 6.7% respectively).

Economic Impact of HEMOS-Cloud Services for M&S Support (M&S 지원을 위한 HEMOS-Cloud 서비스의 경제적 효과)

  • Jung, Dae Yong;Seo, Dong Woo;Hwang, Jae Soon;Park, Sung Uk;Kim, Myung Il
    • KIPS Transactions on Computer and Communication Systems
    • /
    • v.10 no.10
    • /
    • pp.261-268
    • /
    • 2021
  • Cloud computing is a computing paradigm in which users can utilize computing resources in a pay-as-you-go manner. In a cloud system, resources can be dynamically scaled up and down to the user's on-demand so that the total cost of ownership can be reduced. The Modeling and Simulation (M&S) technology is a renowned simulation-based method to obtain engineering analysis and results through CAE software without actual experimental action. In general, M&S technology is utilized in Finite Element Analysis (FEA), Computational Fluid Dynamics (CFD), Multibody dynamics (MBD), and optimization fields. The work procedure through M&S is divided into pre-processing, analysis, and post-processing steps. The pre/post-processing are GPU-intensive job that consists of 3D modeling jobs via CAE software, whereas analysis is CPU or GPU intensive. Because a general-purpose desktop needs plenty of time to analyze complicated 3D models, CAE software requires a high-end CPU and GPU-based workstation that can work fluently. In other words, for executing M&S, it is absolutely required to utilize high-performance computing resources. To mitigate the cost issue from equipping such tremendous computing resources, we propose HEMOS-Cloud service, an integrated cloud and cluster computing environment. The HEMOS-Cloud service provides CAE software and computing resources to users who want to experience M&S in business sectors or academics. In this paper, the economic ripple effect of HEMOS-Cloud service was analyzed by using industry-related analysis. The estimated results of using the experts-guided coefficients are the production inducement effect of KRW 7.4 billion, the value-added effect of KRW 4.1 billion, and the employment-inducing effect of 50 persons per KRW 1 billion.

Changes in Meteorological Variables by SO2 Emissions over East Asia using a Linux-based U.K. Earth System Model (리눅스 기반 U.K. 지구시스템모형을 이용한 동아시아 SO2 배출에 따른 기상장 변화)

  • Youn, Daeok;Song, Hyunggyu;Lee, Johan
    • Journal of the Korean earth science society
    • /
    • v.43 no.1
    • /
    • pp.60-76
    • /
    • 2022
  • This study presents a software full setup and the following test execution times in a Linux cluster for the United Kingdom Earth System Model (UKESM) and then compares the model results from control and experimental simulations of the UKESM relative to various observations. Despite its low resolution, the latest version of the UKESM can simulate tropospheric chemistry-aerosol processes and the stratospheric ozone chemistry using the United Kingdom Chemistry and Aerosol (UKCA) module. The UKESM with UKCA (UKESM-UKCA) can treat atmospheric chemistryaerosol-cloud-radiation interactions throughout the whole atmosphere. In addition to the control UKESM run with the default CMIP5 SO2 emission dataset, an experimental run was conducted to evaluate the aerosol effects on meteorology by changing atmospheric SO2 loading with the newest REAS data over East Asia. The simulation period of the two model runs was 28 years, from January 1, 1982 to December 31, 2009. Spatial distributions of monthly mean aerosol optical depth, 2-m temperature, and precipitation intensity from model simulations and observations over East Asia were compared. The spatial patterns of surface temperature and precipitation from the two model simulations were generally in reasonable agreement with the observations. The simulated ozone concentration and total column ozone also agreed reasonably with the ERA5 reanalyzed one. Comparisons of spatial patterns and linear trends led to the conclusion that the model simulation with the newest SO2 emission dataset over East Asia showed better temporal changes in temperature and precipitation over the western Pacific and inland China. Our results are in line with previous finding that SO2 emissions over East Asia are an important factor for the atmospheric environment and climate change. This study confirms that the UKESM can be installed and operated in a Linux cluster-computing environment. Thus, researchers in various fields would have better access to the UKESM, which can handle the carbon cycle and atmospheric environment on Earth with interactions between the atmosphere, ocean, sea ice, and land.

A Study of Basic Design Method for High Availability Clustering Framework under Distributed Computing Environment (분산컴퓨팅 환경에서의 고가용성 클러스터링 프레임워크 기본설계 연구)

  • Kim, Jeom Goo;Noh, SiChoon
    • Convergence Security Journal
    • /
    • v.13 no.3
    • /
    • pp.17-23
    • /
    • 2013
  • Clustering is required to configure clustering interdependent structural technology. Clustering handles variable workloads or impede continuity of service to continue operating in the event of a failure. Long as high-availability clustering feature focuses on server operating systems. Active-standby state of two systems when the active server fails, all services are running on the standby server, it takes the service. This function switching or switchover is called failover. Long as high-availability clustering feature focuses on server operating systems. The cluster node that is running on multiple systems and services have to duplicate each other so you can keep track of. In the event of a node failure within a few seconds the second node, the node shall perform the duties broken. Structure for high-availability clustering efficiency should be measured. System performance of infrastructure systems performance, latency, response time, CPU load factor(CPU utilization), CPU processes on the system (system process) channels are represented.

Parallel Optimal Power Flow Using PC Clustering (PC 클러스터링을 이용한 병렬 최적조류계산에 관한 연구)

  • Kim, Cheol-Hong;Mun, Kyeong-Jun;Kim, Hyung-Su;Park, J.H.;Kim, Jin-Ho;Lee, Hwa-Seok
    • Proceedings of the KIEE Conference
    • /
    • 2004.11b
    • /
    • pp.190-193
    • /
    • 2004
  • Optimal Power Flow (OPF) is becoming more and more important in the deregulation environment of power pool and there is an urgent need of faster solution technique for on-line application. So this paper presents parallel genetic algorithm-tap search for the solution of the OPF. The control variables modeled unit active power outputs, generator-bus voltage magnitudes and transformer-tap settings. A number of functional operating constraints, such as branch flow limits, load bus boltage magnitude limits and generator reactive capabilities are included as penalties in the fitness function. In parallel GA-TS, GA operators are executed for each process. If best fitness of the GA is not changed for several generations, TS operators are executed for the upper three populations to enhance the local searching capabilities. With migration operation, best string of each node is transferred to the neighboring node after predetermined iterations are executed. For parallel computing, we developed a PC-cluster system consisting of 8 PCs. Each PC employs the 2 GHz Pentium IV CPU and is connected with others through ethernet switch based fast ethernet. To show the usefulness of the proposed method, developed algorithm has been tested and compared on an IEEE 30-bus system in the reference paper. From the simulation results, we can find that the proposed algorithm is efficient for the OPF.

  • PDF

Draft Design of AI Services through Concept Extension of Connected Data Architecture (Connected Data Architecture 개념의 확장을 통한 AI 서비스 초안 설계)

  • Cha, ByungRae;Park, Sun;Oh, Su-Yeol;Kim, JongWon
    • Smart Media Journal
    • /
    • v.7 no.4
    • /
    • pp.30-36
    • /
    • 2018
  • Single domain model like DataLake framework is in spotlight because it can improve data efficiency and process data smarter in big data environment, where large scaled business system generates huge amount of data. In particular, efficient operation of network, storage, and computing resources in logical single domain model is very important for physically partitioned multi-site data process. Based on the advantages of Data Lake framework, we define and extend the concept of Connected Data Architecture and functions of DataLake framework for integrating multiple sites in various domains and managing the lifecycle of data. Also, we propose the design of CDA-based AI service and utilization scenarios in various application domain.

Parallelization of Genome Sequence Data Pre-Processing on Big Data and HPC Framework (빅데이터 및 고성능컴퓨팅 프레임워크를 활용한 유전체 데이터 전처리 과정의 병렬화)

  • Byun, Eun-Kyu;Kwak, Jae-Hyuck;Mun, Jihyeob
    • KIPS Transactions on Computer and Communication Systems
    • /
    • v.8 no.10
    • /
    • pp.231-238
    • /
    • 2019
  • Analyzing next-generation genome sequencing data in a conventional way using single server may take several tens of hours depending on the data size. However, in order to cope with emergency situations where the results need to be known within a few hours, it is required to improve the performance of a single genome analysis. In this paper, we propose a parallelized method for pre-processing genome sequence data which can reduce the analysis time by utilizing the big data technology and the highperformance computing cluster which is connected to the high-speed network and shares the parallel file system. For the reliability of analytical data, we have chosen a strategy to parallelize the existing analytical tools and algorithms to the new environment. Parallelized processing, data distribution, and parallel merging techniques have been developed and performance improvements have been confirmed through experiments.