• Title/Summary/Keyword: Parallel Program

Search Result 586, Processing Time 0.026 seconds

Interprocedural Transformations for Parallel Computing

  • Park, Doo-Soon;Choi, Min-Hyung
    • Journal of Korea Multimedia Society
    • /
    • v.9 no.12
    • /
    • pp.1700-1708
    • /
    • 2006
  • Since the most program execution time is consumed in a loop structure, extracting parallelism from loop programs is critical for the taster program execution. In this paper, we proposed data dependency removal method for a single loop. The data dependency removal method can be applied to uniform and non-uniform data dependency distance in the single loop. Procedure calls parallelisms with only a single loop structure or procedure call most of other methods are concerned with the uniform code within the uniform data dependency distance. We also propose an algorithm, which can be applied to uniform, non-uniform, and complex data dependency distance among the multiple procedures. We compared our method with conventional methods using CRAY-T3E for the performance evaluation. The results show that the proposed algorithm is effective.

  • PDF

Analysis of R410A refrigerant distribution in parallel flow heat exchanger (PF열교환기에서 R410A 냉매분배의 영향)

  • Kim, Jeong-Sik;Kim, Nae-Hyun
    • Proceedings of the SAREK Conference
    • /
    • 2008.06a
    • /
    • pp.340-345
    • /
    • 2008
  • A computer program, which simulates the parall flow evaporator was developed. The program was having used to simulate the sample $650\;mm{\times}190\;mm$ frontal area, 25 mm flow depth and 3.0 mm fin pitch. It was shown that the cooling capacity of 3kW could be available from the sample. The present model, however, does not consider refrigerant mal-distribution in each pass, which is known to reduce the cooling capacity of the parallel flow heat exchanger.

  • PDF

A Parallel Programming Environment Implemented with Graphic User Interface (그래픽 사용자 인터페이스로 구현한 병렬 프로그래밍 환경)

  • Yoo, Jeong-Mok;Lee, Dong-Hee;Lee, Mann-Ho
    • The Transactions of the Korea Information Processing Society
    • /
    • v.7 no.8
    • /
    • pp.2388-2399
    • /
    • 2000
  • This paper describes a parallel programming environment to help programmers to write parallel programs. The parallel programming environment does lexical analysis and syntax analysis like front-end part of common compilers, data flow analysis and data dependence analysis for variables used in programs, and various program transformation methods for parallel programming. Especially, graphic user interface is provided for programmer to get parallel programs easily.

  • PDF

A Study on Effect of Domain-Decomposition Method on Parallel Efficiency in 2-D Flow Computations (2차원 유동장 해석에서 영역분할법에 따른 병렬효율성 검토)

  • Lee Sangyeul;Hur Nahmkeon
    • 한국전산유체공학회:학술대회논문집
    • /
    • 1998.11a
    • /
    • pp.147-152
    • /
    • 1998
  • 2-D flow fields are studied by using a shared memory parallel computer with a parallel flow analysis program which uses domain decomposition method and MPI library for data exchange at overlapped interface. Especially, effects of directional domain decomposition on parallel efficiency are studied for 2-D Lid-Driven cavity flow and flow through square cavity. It is known from the present study that domain decomposition along the main flow direction gives better parallel efficiency in 1-D partitioning than along the other direction. 2-D partitioning, however, is less sensitive to flow directions and gives good parallel efficiency for most of the cases considered.

  • PDF

A Study on the Implementation of GPSS Program on a Parallel Computer (GPSS 프로그램의 병렬화에 관한 연구)

  • 윤정미
    • Journal of the Korea Society for Simulation
    • /
    • v.8 no.2
    • /
    • pp.57-72
    • /
    • 1999
  • With the rapidly increasing complexity of decision-marking or system development in the fields of industry, management, etc., modelling techniques using simulation has become more highlighted. Particularly, the advent of parallel computer systems not only has opened a new horizon of parallel simulation, but also has greatly contributed to the speed-up of the execution of simulation. The implementation of parallel simulation, however, is not a easy job for those who accustomed to the existing computer systems. And it is also necessarily confronted with the problem of synchronization conflict in the process. Thus, how to allow a wider community of users to gain access to parallel simulation while solving synchronization conflicts has become an important issue in simulation study. As a method to solve these problems, this paper is primarily concerned with the implementation of GPSS which is a generally used simulation language for discrete event simulation, onto a parallel computer using C-LINDA. For that, this paper, is to suggest a model and algorithm and to experiment it using a case.

  • PDF

TBBench: A Micro-Benchmark Suite for Intel Threading Building Blocks

  • Marowka, Ami
    • Journal of Information Processing Systems
    • /
    • v.8 no.2
    • /
    • pp.331-346
    • /
    • 2012
  • Task-based programming is becoming the state-of-the-art method of choice for extracting the desired performance from multi-core chips. It expresses a program in terms of lightweight logical tasks rather than heavyweight threads. Intel Threading Building Blocks (TBB) is a task-based parallel programming paradigm for multi-core processors. The performance gain of this paradigm depends to a great extent on the efficiency of its parallel constructs. The parallel overheads incurred by parallel constructs determine the ability for creating large-scale parallel programs, especially in the case of fine-grain parallelism. This paper presents a study of TBB parallelization overheads. For this purpose, a TBB micro-benchmarks suite called TBBench has been developed. We use TBBench to evaluate the parallelization overheads of TBB on different multi-core machines and different compilers. We report in detail in this paper on the relative overheads and analyze the running results.

Direct Numerical Simulation of Active Fiber Composite (능동 섬유 복합재의 직접적 수치 모사)

  • 백승훈;김승조
    • Proceedings of the Korean Society For Composite Materials Conference
    • /
    • 2003.04a
    • /
    • pp.5-9
    • /
    • 2003
  • Stress and deflection of Active Fiber Composite(AFC) embedded and/or attached composite structures are numerically investigated at the constituent level by the Direct Numerical Simulation(DNS). The DNS approach which models and simulates the fiber and matrix directly using 3D finite elements need to be solved by efficient way. To handle this large scale problem, parallel program for solving piezoelectric behavior was developed and run on the parallel computing environment. Also, the stress result from DNS approach is compared with that from uniform field model.

  • PDF

Distributed/parallel Algorithm Simulator (분산 및 병렬 알고리즘 시뮬레이터)

  • ;R.S.Ramakrishna
    • Proceedings of the Korean Information Science Society Conference
    • /
    • 1999.10c
    • /
    • pp.777-779
    • /
    • 1999
  • A new distributed/parallel algorithm simulator, DASim(Distributed Algorithm Simulator), is proposed in this paper. The idea is to ease the task of design, analysis and implementation of distributed algorithms. A small high level language has been proposed for the purpose. Through this non-language specific high level language, the users are spared from the tedious details about how to program distributed or parallel algorithms. Further, visualization of these algorithms are pretty helpful to understand behaviors of these algorithms.

  • PDF

Space-Sharing Scheduling Schemes for NOW with Heterogeneous Computing Power (이질적 계산 능력을 가진 NOW를 위한 공간 공유 스케쥴링 기법)

  • Kim, Jin-Sung;Shim, Young-Chul
    • Journal of KIISE:Computer Systems and Theory
    • /
    • v.27 no.7
    • /
    • pp.650-664
    • /
    • 2000
  • NOW(Network of Workstations) is considered as a platform for running parallel programs by many people. One of the fundamental problems that must be addressed to achieve good performance for parallel programs on NOW is the determination of efficient job scheduling policies. Currently most research on NOW assumes that all the workstations in the NOW have the same processing power. In this paper we consider a NOW in which workstations may have different computing power. We introduce 10 classes of space sharing-based scheduling policies that can be applied to the NOW with heterogeneous computing power. We compare the performance of these scheduling policies by using the simulator which accepts synthetically generated sequential and parallel workloads and generates the response time and waiting time of parallel jobs as performance indices of various scheduling strategies. Through the experiments the case when a parallel program is partitioned heterogeneously in proportion to the computing power of workstations is shown to have better performance than when a parallel program is partitioned into parallel processes of the same size. When the owner returns to the workstation which is executing a parallel process, the policy which just lowers the priority of the parallel process shows better performance than the one which migrates the parallel process to a new idle workstation. Among the policies which use heterogeneous partitioning and process priority lowering, the adaptive policy performed best across the wide range of inter-arrival time of parallel programs but when the load imbalance among parallel processes becomes very high, the modified adaptive policy performed better.

  • PDF

A Study on Efficient Executions of MPI Parallel Programs in Memory-Centric Computer Architecture

  • Lee, Je-Man;Lee, Seung-Chul;Shin, Dongha
    • Journal of the Korea Society of Computer and Information
    • /
    • v.25 no.1
    • /
    • pp.1-11
    • /
    • 2020
  • In this paper, we present a technique that executes MPI parallel programs, that are developed on processor-centric computer architecture, more efficiently on memory-centric computer architecture without program modification. The technique we present here improves performance by replacing low-speed data communication over the network of MPI library functions with high-speed data communication using the property called fast large shared memory of memory-centric computer architecture. The technique we present in the paper is implemented in two programs. The first program is a modified MPI library called MC-MPI-LIB that runs MPI parallel programs more efficiently on memory-centric computer architecture preserving the semantics of MPI library functions. The second program is a simulation program called MC-MPI-SIM that simulates the performance of memory-centric computer architecture on processor-centric computer architecture. We developed and tested the programs on distributed systems environment deployed on Docker based virtualization. We analyzed the performance of several MPI parallel programs and showed that we achieved better performance on memory-centric computer architecture. Especially we could see very high performance on the MPI parallel programs with high communication overhead.