• Title/Summary/Keyword: Distributed and Parallel Algorithms

Search Result 77, Processing Time 0.025 seconds

Distributed/parallel Algorithm Simulator (분산 및 병렬 알고리즘 시뮬레이터)

  • ;R.S.Ramakrishna
    • Proceedings of the Korean Information Science Society Conference
    • /
    • 1999.10c
    • /
    • pp.777-779
    • /
    • 1999
  • A new distributed/parallel algorithm simulator, DASim(Distributed Algorithm Simulator), is proposed in this paper. The idea is to ease the task of design, analysis and implementation of distributed algorithms. A small high level language has been proposed for the purpose. Through this non-language specific high level language, the users are spared from the tedious details about how to program distributed or parallel algorithms. Further, visualization of these algorithms are pretty helpful to understand behaviors of these algorithms.

  • PDF

Distributed Genetic Algorithms for the TSP (분산 유전알고리즘의 TSP 적용)

  • 박유석
    • Journal of the Korea Safety Management & Science
    • /
    • v.3 no.3
    • /
    • pp.191-200
    • /
    • 2001
  • Parallel Genetic Algorithms partition the whole population into several sub-populations and search the optimal solution by exchanging the information each others periodically. Distributed Genetic Algorithm, one of Parallel Genetic Algorithms, divides a large population into several sub-populations and executes the traditional Genetic Algorithm on each sub-population independently. And periodically promising individuals selected from sub-populations are migrated by following the migration interval and migration rate to different sub-populations. In this paper, for the Travelling Salesman Problems, we analyze and compare with Distributed Genetic Algorithms using different Genetic Algorithms and using same Genetic Algorithms on each separated sub-population The simulation result shows that using different Genetic Algorithms obtains better results than using same Genetic Algorithms in Distributed Genetic Algorithms. This results look like the property of rapidly searching the approximated optima and keeping the variety of solution make interaction in different Genetic Algorithms.

  • PDF

A Development of Distributed Parallel Processing algorithm for Power Flow analysis (전력 조류 계산의 분산 병렬처리기법에 관한 연구)

  • Lee, Chun-Mo;Lee, Hae-Ki
    • Proceedings of the KIEE Conference
    • /
    • 2001.07e
    • /
    • pp.134-140
    • /
    • 2001
  • Parallel processing has the potential to be cost effectively used on computationally intense power system problems. But this technology is not still available is not only parallel computer but also parallel processing scheme. Testing these algorithms to ensure accuracy, and evaluation of their performance is also an issue. Although a significant amount of parallel algorithms of power system problem have been developed in last decade, actual testing on processor architectures lies in the beginning stages. This paper presents the parallel processing algorithm to supply the base being able to treat power flow by newton's method by the distributed memory type parallel computer. This method is to assign and to compute teared blocks of sparse matrix at each parallel processors. The testing to insure accuracy of developed method have been done on serial computer by trying to simulate a parallel environment.

  • PDF

Proposition and Evaluation of Parallelism-Independent Scheduling Algorithms for DAGs of Tasks with Non-Uniform Execution Time

  • Kirilka Nikolova;Atusi Maeda;Sowa, Masa-Hiro
    • Proceedings of the IEEK Conference
    • /
    • 2000.07a
    • /
    • pp.289-293
    • /
    • 2000
  • We propose two new algorithms for parallelism-independent scheduling. The machine code generated from the compiler using these algorithms in its scheduling phase is parallelism-independent code, executable in minimum time regardless of the number of the processors in the parallel computer. Our new algorithms have the following phases: finding the minimum number of processors on which the program can be executed in minimal time, scheduling by an heuristic algorithm for this predefined number of processors, and serialization of the parallel schedule according to the earliest start time of the tasks. At run time tasks are taken from the serialized schedule and assigned to the processor which allows the earliest start time of the task. The order of the tasks decided at compile time is not changed at run time regardless of the number of the available processors which means there is no out-of-order issue and execution. The scheduling is done predominantly at compile time and dynamic scheduling is minimized and diminished to allocation of the tasks to the processors. We evaluate the proposed algorithms by comparing them in terms of schedule length to the CP/MISF algorithm. For performance evaluation we use both randomly generated DAGs (directed acyclic graphs) and DACs representing real applications. From practical point of view, the algorithms we propose can be successfully used for scheduling programs for in-order superscalar processors and shared memory multiprocessor systems. Superscalar processors with any number of functional units can execute the parallelism-independent code in minimum time without necessity for dynamic scheduling and out-of-order issue hardware. This means that the use of our algorithms will lead to reducing the complexity of the hardware of the processors and the run-time overhead related to the dynamic scheduling.

  • PDF

Performance Optimization of Parallel Algorithms

  • Hudik, Martin;Hodon, Michal
    • Journal of Communications and Networks
    • /
    • v.16 no.4
    • /
    • pp.436-446
    • /
    • 2014
  • The high intensity of research and modeling in fields of mathematics, physics, biology and chemistry requires new computing resources. For the big computational complexity of such tasks computing time is large and costly. The most efficient way to increase efficiency is to adopt parallel principles. Purpose of this paper is to present the issue of parallel computing with emphasis on the analysis of parallel systems, the impact of communication delays on their efficiency and on overall execution time. Paper focuses is on finite algorithms for solving systems of linear equations, namely the matrix manipulation (Gauss elimination method, GEM). Algorithms are designed for architectures with shared memory (open multiprocessing, openMP), distributed-memory (message passing interface, MPI) and for their combination (MPI + openMP). The properties of the algorithms were analytically determined and they were experimentally verified. The conclusions are drawn for theory and practice.

Applying Distributed Agents to Parallel Genetic Algorithm on Dynamic Network Environments (동적 네트워크 환경하의 분산 에이전트를 활용한 병렬 유전자 알고리즘 기법)

  • Baek Jin-Wook;Bang Jeon-Won
    • Journal of the Korea Society of Computer and Information
    • /
    • v.11 no.4 s.42
    • /
    • pp.119-125
    • /
    • 2006
  • Distributed Systems can be defined as set of computing resources connected by computer network. One of the most significant techniques in optimization problem domains is parallel genetic algorithms, which are based on distributed systems. Since the status of dynamic network environments such as Internet and mobile computing. can be changed continually, it must not be efficient on the dynamic environments to solve an optimization problem using previous parallel genetic algorithms themselves. In this paper, we propose the effective technique, in which the parallel genetic algorithm can be used efficiently on the dynamic network environments.

  • PDF

Performance Evaluation of a Parallel DEVS Simulation Environment of P-DEVSIM ++ (병렬 DEVS 시뮬레이션 환경(P-DEVSIM ++) 성능 평가)

  • 성영락
    • Journal of the Korea Society for Simulation
    • /
    • v.2 no.1
    • /
    • pp.31-44
    • /
    • 1993
  • Zeigler's DEVS(Discrete Event Systems Specification) formalism supports formal specification of discrete event systems in a hierarchical , modular manner. Associated are hierarchical, distributed simulation algorithms, called abstract simulators, which interpret dynamics of DEVS models. This paper deals with performance evaluation of P-DEVSIM ++, a parallel simulation environment which implements the DEVS formalism and associated simulation algorithms in a parallel environment. Performance simulator has been developed and used to experiment models of parallel simulation executions in different conditions. The experimental result shows that simulation time depends on both the number of processors in the parallel system and the communication overheads among such processors.

  • PDF

Distributed Structural Analysis Algorithms for Large-Scale Structures based on PCG Algorithms (대형구조물의 분산구조해석을 위한 PCG 알고리즘)

  • 권윤한;박효선
    • Journal of the Computational Structural Engineering Institute of Korea
    • /
    • v.12 no.3
    • /
    • pp.385-396
    • /
    • 1999
  • In the process of structural design for large-scale structures with several thousands of degrees of freedom, a plethora of structural calculations with large amount of data storage are required to obtain the forces and displacements of the members. However, current computational environment with single microprocessor such as a personal computer or a workstation is not capable of generating a high-level of efficiency in structural analysis and design process for large-scale structures. In this paper, a high-performance parallel computing system interconnected by a network of personal computers is proposed for an efficient structural analysis. Two distributed structural analysis algorithms are developed in the form of distributed or parallel preconditioned conjugate gradient (DPCG) method. To enhance the performance of the developed distributed structural analysis algorithms, the number of communications and the size of data to be communicated are minimized. These algorithms are applied to the structural analyses of three large space structures as well as a 144-story tube-in-tube framed structure.

  • PDF

A Study on Sorting in A Computer Using The Binary Multi-level Multi-access Protocol

  • Jung Chang-Duk
    • Proceedings of the Korea Inteligent Information System Society Conference
    • /
    • 2006.06a
    • /
    • pp.303-310
    • /
    • 2006
  • The sorting algorithms have been developed to take advantage of distributed computers. But the speedup of parallel sorting algorithms decrease rapidly with increased number of processors due to parallel processing overhead such as context switching time and inter-processor communication cost. In this paper, we propose a parallel sorting method which provides linear speedup of an optimal serial algorithm for a system with a large number of processors. This algorithm may even provide superlinear speedup for a practical system. The algorithm takes advantage of an interconnection network properties and its protocol.

  • PDF

Fully Distributed Economic Dispatching Methods Based on Alternating Direction Multiplier Method

  • Yang, Linfeng;Zhang, Tingting;Chen, Guo;Zhang, Zhenrong;Luo, Jiangyao;Pan, Shanshan
    • Journal of Electrical Engineering and Technology
    • /
    • v.13 no.5
    • /
    • pp.1778-1790
    • /
    • 2018
  • Based on the requirements and characteristics of multi-zone autonomous decision-making in modern power system, fully distributed computing methods are needed to optimize the economic dispatch (ED) problem coordination of multi-regional power system on the basis of constructing decomposition and interaction mechanism. In this paper, four fully distributed methods based on alternating direction method of multipliers (ADMM) are used for solving the ED problem in distributed manner. By duplicating variables, the 2-block classical ADMM can be directly used to solve ED problem fully distributed. The second method is employing ADMM to solve the dual problem of ED in fully distributed manner. N-block methods based on ADMM including Alternating Direction Method with Gaussian back substitution (ADM_G) and Exchange ADMM (E_ADMM) are employed also. These two methods all can solve ED problem in distributed manner. However, the former one cannot be carried out in parallel. In this paper, four fully distributed methods solve the ED problem in distributed collaborative manner. And we also discussed the difference of four algorithms from the aspects of algorithm convergence, calculation speed and parameter change. Some simulation results are reported to test the performance of these distributed algorithms in serial and parallel.