• Title/Summary/Keyword: message passing interface (MPI)

Search Result 115, Processing Time 0.03 seconds

Parallelization of sheet forming analysis program using MPI (MPI를 이용한 판재성형해석 프로그램의 병렬화)

  • Kim, Eui-Joong;Suh, Yeong-Sung
    • Transactions of the Korean Society of Mechanical Engineers A
    • /
    • v.22 no.1
    • /
    • pp.132-141
    • /
    • 1998
  • A parallel version of sheet forming analysis program was developed. This version is compatible with any parallel computers which support MPI that is one of the most recent and popular message passing libraries. For this purpose, SERI-SFA, a vector version which runs on Cray Y-MP C90, a sequential vector computer, was used as a source code. For the sake of the effectiveness of the work, the parallelization was focused on the selected part after checking the rank of CPU consumed from the exemplary calculation on Cray Y-MP C90. The subroutines associated with contact algorithm was selected as targe parts. For this work, MPI was used as a message passing library. For the performance verification, an oil pan and an S-rail forming simulation were carried out. The performance check was carried out by the kernel and total CPU time along with theoretical performance using Amdahl's Law. The results showed some performance improvement within the limit of the selective paralellization.

Design of Thin-Client Framework for Application Sharing & Optimization of Data Access (애플리케이션 공유 및 데이터 접근 최적화를 위한 씬-클라이언트 프레임워크 설계)

  • Song, Min-Gyu
    • Journal of Korea Society of Industrial Information Systems
    • /
    • v.14 no.5
    • /
    • pp.19-32
    • /
    • 2009
  • In this paper, we design thin-client framework capable of application sharing & data access on the Internet, and apply related skills, such as X windows system, pseudo server, CODA file system, MPI(Message Passing Interface). We suggest a framework for the thin client to access data produced by working on a server optimally as well as to run server side application, even in the case of network down. Additionally, it needed to reflect all local computing changes to remote server when network is restored. To design thin client framework with these characteristics, in this paper, we apply distributed pseudo server and CODA file system to our framework, also utilize MPI for the purpose of more efficient computing & management. It allows for implementation of network independent computing environment of thin client, also provide scalable application service to numerous user through the elimination of bottleneck on caused by server overload. In this paper, we discuss the implementing method of thin client framework in detail.

MPI-GWAS: a supercomputing-aided permutation approach for genome-wide association studies

  • Paik, Hyojung;Cho, Yongseong;Cho, Seong Beom;Kwon, Oh-Kyoung
    • Genomics & Informatics
    • /
    • v.20 no.1
    • /
    • pp.14.1-14.4
    • /
    • 2022
  • Permutation testing is a robust and popular approach for significance testing in genomic research that has the advantage of reducing inflated type 1 error rates; however, its computational cost is notorious in genome-wide association studies (GWAS). Here, we developed a supercomputing-aided approach to accelerate the permutation testing for GWAS, based on the message-passing interface (MPI) on parallel computing architecture. Our application, called MPI-GWAS, conducts MPI-based permutation testing using a parallel computing approach with our supercomputing system, Nurion (8,305 compute nodes, and 563,740 central processing units [CPUs]). For 107 permutations of one locus in MPI-GWAS, it was calculated in 600 s using 2,720 CPU cores. For 107 permutations of ~30,000-50,000 loci in over 7,000 subjects, the total elapsed time was ~4 days in the Nurion supercomputer. Thus, MPI-GWAS enables us to feasibly compute the permutation-based GWAS within a reason-able time by harnessing the power of parallel computing resources.

Parallel VHDL Simulation on IBM SP2 and SGI Origin 2000 (IBM SP2와 SGI Origin 2000에서의 병렬 VHDL 시뮬레이션)

  • 정영식
    • Journal of the Korea Society for Simulation
    • /
    • v.7 no.1
    • /
    • pp.69-83
    • /
    • 1998
  • In this paper, we present the results of simulation by running parallel VHDL simulation on typical MPP(Massively Parallel Processor) systems such as IBM SP2 and SGI Origin 2000. Parallel simulation uses the synchronous protocol and parallel program is implemented using MPI(Message Passing Interface) based on message passing model, so that it can urn on any parallel programming environment which supports MPI, a standard communication library. And then GVT(Global Virtual Time) computation for parallel simulation is based on the global broadcasting with MPI_Bcast(), which is a standard function in MPI and piggybacking. Our benchmark exhibits that as size of VHDL grows, the parallel simulation has a better performance compared with the sequential simulation. In addition, we also show the results of comparison between IBM SP2 and SGI Origin 2000 by applying the same application to those indirectly.

  • PDF

Large Eddy Simulation of Turbulent Flow around a Ship Model Using Message Passing Interface (병렬계산기법을 이용한 선체주위 점성유동장의 LES해석)

  • Choi, Hee-Jong;Yoon, Hyun-Sik;Chun, Ho-Hwan;Kang, Dae-Hwan;Park, Jong-Chun
    • Journal of Ocean Engineering and Technology
    • /
    • v.20 no.4 s.71
    • /
    • pp.76-82
    • /
    • 2006
  • The large-eddy simulation(LES) technique, based an a message passing interface method(MPI), was applied to investigate the turbulent flaw phenomena around a ship. The Smagorinski model was used in the present LES simulation to simulate the turbulent flaw around a ship. The SPMD(sidsngle program multiple data) technique was used for parallelization of the program using MPI. All computations were performed an a 24-node PC cluster parallel machine, composed of 2.6 GHz CPU, which had been installed in the Advanced Ship Engineering Research Center(ASERC). Numerical simulations were performed for the Wigley hull, and the Series 60 hull(CB=0.6) using 1/4-, 1/2-, 1- and 2-million grid systems and the computational results had been compared to the experimental ones.

Parallel Processing of Airborne Laser Scanning Data Using a Hybrid Model Based on MPI and OpenMP (MPI와 OpenMP기반 하이브리드 모델을 이용한 항공 레이저 스캐닝 자료의 병렬 처리)

  • Han, Soo-Hee;Park, Il-Suk;Heo, Joon
    • Journal of the Korean Society of Surveying, Geodesy, Photogrammetry and Cartography
    • /
    • v.30 no.2
    • /
    • pp.135-142
    • /
    • 2012
  • In the present study, a parallel processing method running on a multi-core PC-Cluster is introduced to produce digital surface model (DSM) and digital terrain model (DTM) from huge airborne laser scanning data. A hybrid model using both message passing interface (MPI) and OpenMP was devised by revising a conventional MPI model which utilizes only MPI, and tested on a multi-core PC-Cluster for performance validation. In the results, the hybrid model has not shown better performances in the interpolation process to produce DSM, but the overall performance has turned out to be better by the help of reduced MPI calls. Additionally, scheduling function of OpenMP has revealed its ability to enhance the performance by controlling inequal overloads charged on cores induced by irregular distribution of airborne laser scanning data.

The Mixed Finite Element Analysis for Nearly Incompressible and Impermeable Porous Media Using Parallel Algorithm (병렬알고리즘 이용한 비압축, 비투과성 포화 다공질매체의 혼합유한요소해석)

  • Tak, Moon-Ho;Kang, Yoon-Sik;Park, Tae-Hyo
    • Journal of the Computational Structural Engineering Institute of Korea
    • /
    • v.23 no.4
    • /
    • pp.361-368
    • /
    • 2010
  • In this paper, the parallel algorithm using MPI(Message-Passing Interface) library is introduced in order to improve numerical efficiency for the staggered method for nearly incompressible and impermeable porous media which was introduced by Park and Tak(2010). The porous media theory and the staggered method are also briefly introduced in this paper. Moreover, we account for MPI library for blocking, non-blocking, and collective communication, and propose combined the staggered method with the blocking and nonblocking MPI library. And then, we present how to allocate CPUs on the staggered method and the MPI library, which is related with the numerical efficiency in order to solve unknown variables on nearly incompressible and impermeable porous media. Finally, the results comparing serial solution with parallel solution are verified by 2 dimensional saturated porous model according to the number of FEM meshes.

The Mixed Finite Element Analysis for Saturated Porous Media using FETI Method

  • Lee, Kyung-Jae;Tak, Moon-Ho;Park, Tae-Hyo
    • Journal of the Computational Structural Engineering Institute of Korea
    • /
    • v.23 no.6
    • /
    • pp.693-702
    • /
    • 2010
  • In this paper, FETI(Finite Element Tearing and Interconnecting) method is introduced in order to improve numerical efficiency of Staggered method. The porous media theory, the Staggered method and the FETI method are briefly introduced in this paper. In addition, we account for the MPI(Message Passing Interface) library for parallel analysis, and the proposed combined Staggered method with FETI method. Finally Lagrange multipliers and CG(Conjugate Gradient) algorithm to solve decomposed domain are proposed, and then the proposed method is verified to be numerically efficient by MPI library.

A Dynamic Co-scheduling Scheme for MPI-based Parallel Programs on Linux Clusters (리눅스 클러스터에서 MPI 기반 병렬 프로그램의 동적 동시 스케줄링 기법)

  • Kim, Hyuk;Rhee, Yun-Seok
    • Journal of the Korea Society of Computer and Information
    • /
    • v.13 no.1
    • /
    • pp.29-35
    • /
    • 2008
  • For efficient message passing of Parallel programs, it is required to schedule the involved two processes at the same time which are executed on different nodes, that is called 'co-scheduling' However, each node of cluster systems is built on top of general purpose multitasking OS. which autonomously manages local Processes. Thus it is not so easy to co-schedule two (or more) processes in such computing environment. Our work proposes a co-scheduling scheme for MPI-based parallel programs which exploits message exchange information between two parties. We implement the scheme on Linux cluster which requires slight kernel hacking and MPI library modification. The experiment with NPB parallel suite shows that our scheme results in 33-56% reduction in the execution time compared to the typical scheduling case. and especially better Performance in more communication-bound applications.

  • PDF

A Fault-Tolerant Linear System Solver in a Standard MPI Environment (표준 MPI 환경에서의 무정지형 선형 시스템 해법)

  • Park, Pil-Seong
    • Journal of Internet Computing and Services
    • /
    • v.6 no.6
    • /
    • pp.23-34
    • /
    • 2005
  • In a large scale parallel computation, failures of some nodes or communication links end up with waste of computing resources, Several fault-tolerant MPI libraries have been proposed so far, but the programs written by using such libraries have a portability problem since fault-tolerant features are not supported by the MPI standard yet, In this paper, we propose an application-level fault-tolerant linear system solver that uses the asynchronous iteration algorithm and the standard MPI functions only, which does not have a portability problem and is more efficient by adopting a simplified recovery mechanism.

  • PDF