• Title/Summary/Keyword: High Performance Computing

Search Result 1,099, Processing Time 0.03 seconds

A Fundamental Study of Thermal-Fluid Flow Analysis using High Performance Computing under the GRID (그리드 환경하에서 고성능 컴퓨팅을 이용한 열유동 해석 기법에 관한 기초연구)

  • Hong, Seung-Do;Lee, Dae-Sung;Lee, Jae-Ryong;Ha, Man-Yeong;Lee, Sang-San
    • Proceedings of the KSME Conference
    • /
    • 2003.11a
    • /
    • pp.928-933
    • /
    • 2003
  • For simulation of three-dimensional turbulent flow with LES and DNS takes much time and expense with current available computing resources. It is nearly impossible to simulate turbulent flow with high Reynolds number. So, the emerging alternative is the Grid computing for needed computation power and working environment. In this study, the CFD code was parallelized to adapt it for the parallel computing under the Grid environment. In the first place, the Grid environment was built to connect the PC-Cluster facilities belong to the different institutions using communication network system. And CFD applications were calculated to check the performance of the parallel code developed for the Grid environment. Although it is a fundamental study, it brings about a important meaning as first step in research of the Grid.

  • PDF

Numerical study of dynamic buckling for plate and shell structures

  • Liu, Z.S.;Lee, H.P.;Lu, C.
    • Structural Engineering and Mechanics
    • /
    • v.20 no.2
    • /
    • pp.241-257
    • /
    • 2005
  • A numerical approach combining the finite element method with two different stability criteria namely the Budiansky and the phase-plane buckling criteria is used to study the dynamic buckling phenomena of plate and shell structures subjected to sudden applied loading. In the finite element analysis an explicit time integration scheme is used and the two criteria are implemented in the Finite Element analysis. The dynamic responses of the plate and shell structures have been investigated for different values of the plate and shell imperfection factors. The results indicate that the dynamic buckling time, which is normally considered in predicting elasto-plastic buckling behavior, should be taken into consideration with the buckling criteria for elastic buckling analysis of plate and shell structures. By selecting proper control variables and incorporating them with two dynamic buckling criteria, the unique dynamic buckling load can be obtained and the problems of ambiguity and contradiction of dynamic buckling load of plate and shell structure can be resolved.

Development of a meshless finite mixture (MFM) method

  • Cheng, J.Q.;Lee, H.P.;Li, Hua
    • Structural Engineering and Mechanics
    • /
    • v.17 no.5
    • /
    • pp.671-690
    • /
    • 2004
  • A meshless method with novel variation of point collocation by finite mixture approximation is developed in this paper, termed the meshless finite mixture (MFM) method. It is based on the finite mixture theorem and consists of two or more existing meshless techniques for exploitation of their respective merits for the numerical solution of partial differential boundary value (PDBV) problems. In this representation, the classical reproducing kernel particle and differential quadrature techniques are mixed in a point collocation framework. The least-square method is used to optimize the value of the weight coefficient to construct the final finite mixture approximation with higher accuracy and numerical stability. In order to validate the developed MFM method, several one- and two-dimensional PDBV problems are studied with different mixed boundary conditions. From the numerical results, it is observed that the optimized MFM weight coefficient can improve significantly the numerical stability and accuracy of the newly developed MFM method for the various PDBV problems.

Bandwidth-aware Memory Placement on Hybrid Memories targeting High Performance Computing Systems

  • Lee, Jongmin
    • Journal of the Korea Society of Computer and Information
    • /
    • v.24 no.8
    • /
    • pp.1-8
    • /
    • 2019
  • Modern computers provide tremendous computing capability and a large memory system. Hybrid memories consist of next generation memory devices and are adopted in high performance systems. However, the increased complexity of the microprocessor makes it difficult to operate the system effectively. In this paper, we propose a simple data migration method called Bandwidth-aware Data Migration (BDM) to efficiently use memory systems for high performance processors with hybrid memory. BDM monitors the status of applications running on the system using hardware performance monitoring tools and migrates the appropriate pages of selected applications to High Bandwidth Memory (HBM). BDM selects applications whose bandwidth usages are high and also evenly distributed among the threads. Experimental results show that BDM improves execution time by an average of 20% over baseline execution.

A Novel Smart Contract based Optimized Cloud Selection Framework for Efficient Multi-Party Computation

  • Haotian Chen;Abir EL Azzaoui;Sekione Reward Jeremiah;Jong Hyuk Park
    • Journal of Information Processing Systems
    • /
    • v.19 no.2
    • /
    • pp.240-257
    • /
    • 2023
  • The industrial Internet of Things (IIoT) is characterized by intelligent connection, real-time data processing, collaborative monitoring, and automatic information processing. The heterogeneous IIoT devices require a high data rate, high reliability, high coverage, and low delay, thus posing a significant challenge to information security. High-performance edge and cloud servers are a good backup solution for IIoT devices with limited capabilities. However, privacy leakage and network attack cases may occur in heterogeneous IIoT environments. Cloud-based multi-party computing is a reliable privacy-protecting technology that encourages multiparty participation in joint computing without privacy disclosure. However, the default cloud selection method does not meet the heterogeneous IIoT requirements. The server can be dishonest, significantly increasing the probability of multi-party computation failure or inefficiency. This paper proposes a blockchain and smart contract-based optimized cloud node selection framework. Different participants choose the best server that meets their performance demands, considering the communication delay. Smart contracts provide a progressive request mechanism to increase participation. The simulation results show that our framework improves overall multi-party computing efficiency by up to 44.73%.

Traffic Engineering with Segment Routing under Uncertain Failures

  • Zheng, Zengwei;Zhao, Chenwei;Zhang, Jianwei;Cai, Jianping
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.15 no.7
    • /
    • pp.2589-2609
    • /
    • 2021
  • Segment routing (SR) is a highly implementable approach for traffic engineering (TE) with high flexibility, high scalability, and high stability, which can be established upon existing network infrastructure. Thus, when a network failure occurs, it can leverage the existing rerouting methods, such as rerouting based on Interior Gateway Protocol (IGP) and fast rerouting with loop-free alternates. To better exploit these features, we propose a high-performance and easy-to-deploy method SRUF (Segment Routing under Uncertain Failures). The method is inspired by the Value-at-Risk (VaR) theory in finance. Just as each investment risk is considered in financial investment, SRUF also considers each traffic distribution scheme's risk when forwarding traffic to achieve optimal traffic distribution. Specifically, SRUF takes into account that every link may fail and therefore has inherent robustness and high availability. Also, SRUF considers that a single link failure is a low-probability event; hence it can achieve high performance. We perform experiments on real topologies to validate the flexibility, high-availability, and load balancing of SRUF. The results show that when given an availability requirement, SRUF has greater load balancing performance under uncertain failures and that when given a demand requirement, SRUF can achieve higher availability.

The Analysis of the Supercomputer Trends in Weather and Climate Research Areas (기상 및 기후 연구 분야의 슈퍼컴퓨터 보유 추이 분석)

  • Joh, Minsu;Park, Hyei-Sun
    • Atmosphere
    • /
    • v.15 no.2
    • /
    • pp.119-127
    • /
    • 2005
  • It is challenging work to predict weather and climate conditions of the future in advance. Since ENIAC was developed, weather and climate research areas have been taking advantage of the improvements in computer hardware. High performance computers allows researchers to build high quality models that allow them to make good predictions of what might happen in the future. Statistics on the high performance computers are one of the major interest to not only manufacturers but also the users such as weather and climate researchers. For this reason, the Top500 Supercomputer Sites Report has been being released twice a year since 1993 to provide a reliable basis for tracking and detecting trends in high performance computing. Using the Top500 Report, a short review on the supercomputer trends in weather and climate research areas is provided in this article.

Parallel LDPC Decoder for CMMB on CPU and GPU Using OpenCL (OpenCL을 활용한 CPU와 GPU 에서의 CMMB LDPC 복호기 병렬화)

  • Park, Joo-Yul;Hong, Jung-Hyun;Chung, Ki-Seok
    • IEMEK Journal of Embedded Systems and Applications
    • /
    • v.11 no.6
    • /
    • pp.325-334
    • /
    • 2016
  • Recently, Open Computing Language (OpenCL) has been proposed to provide a framework that supports heterogeneous computing platforms. By using an OpenCL framework, digital communication systems can support various protocols in a unified computing environment to achieve both high portability and high performance. This article introduces a parallel software decoder of Low Density Parity Check (LDPC) codes for China Multimedia Mobile Broadcasting (CMMB) on a heterogeneous platform. Each step of LDPC decoding has different parallelization characteristics. In this paper, steps suitable for task-level parallelization are executed on the CPU, and steps suitable for data-level parallelization are processed by the GPU. To improve the performance of the proposed OpenCL kernels for LDPC decoding operations, explicit thread scheduling, loop-unrolling, and effective data transfer techniques are applied. The proposed LDPC decoder achieves high performance by using heterogeneous multi-core processors on a unified computing framework.

On Effective Slack Reclamation in Task Scheduling for Energy Reduction

  • Lee, Young-Choon;Zomaya, Albert Y.
    • Journal of Information Processing Systems
    • /
    • v.5 no.4
    • /
    • pp.175-186
    • /
    • 2009
  • Power consumed by modern computer systems, particularly servers in data centers has almost reached an unacceptable level. However, their energy consumption is often not justifiable when their utilization is considered; that is, they tend to consume more energy than needed for their computing related jobs. Task scheduling in distributed computing systems (DCSs) can play a crucial role in increasing utilization; this will lead to the reduction in energy consumption. In this paper, we address the problem of scheduling precedence-constrained parallel applications in DCSs, and present two energy- conscious scheduling algorithms. Our scheduling algorithms adopt dynamic voltage and frequency scaling (DVFS) to minimize energy consumption. DVFS, as an efficient power management technology, has been increasingly integrated into many recent commodity processors. DVFS enables these processors to operate with different voltage supply levels at the expense of sacrificing clock frequencies. In the context of scheduling, this multiple voltage facility implies that there is a trade-off between the quality of schedules and energy consumption. Our algorithms effectively balance these two performance goals using a novel objective function and its variant, which take into account both goals; this claim is verified by the results obtained from our extensive comparative evaluation study.

A Study on Construction Site of Virtual Desktop Infrastructure (VDI) System Model for Cloud Computing BIM Service

  • Lee, K.H.;Kwon, S.W.;Shin, J.H.;Choi, G.S.;Moon, D.Y.
    • International conference on construction engineering and project management
    • /
    • 2015.10a
    • /
    • pp.665-666
    • /
    • 2015
  • Recently BIM technology has been expanded for using in construction project. However its spread has been delayed than the initial expectations, due to the high-cost of BIM infrastructure development, the lack of regulations, the lack of process and so forth. In construction site phase, especially the analysis of current research trend about IT technologies, virtualization and BIM service, data exchange such as drawing, 3D model, object data, properties using cloud computing and virtual server system is defined as a most successful solution. The purpose of this study is enable the cloud computing BIM server to provide several main function such as edit a model, 3D model viewer and checker, mark-up, snapshot in high-performance quality by proper design of VDI system. Concurrent client connection performance is a main technical index of VDI. Through test-bed server client, developed VDI system's multi-connect control will be evaluated. The performance-test result of BIM server VDI will effect to development direction of cloud computing BIM service for commercialization.

  • PDF