• Title/Summary/Keyword: Graphics Processing Unit : GPU

Search Result 154, Processing Time 0.024 seconds

Performance Improvement of Web Service Based on GPGPU and Task Queue

  • Kim, Changsu;Kim, Kyunghwan;Jung, Hoekyung
    • Journal of information and communication convergence engineering
    • /
    • v.19 no.4
    • /
    • pp.257-262
    • /
    • 2021
  • Providing web services to users has become expensive in recent times. For better web services, a web server is provided with high-performance technology. To achieve great web service experiences, tools such as general-purpose graphics processing units (GPGPUs), artificial intelligence, high-performance computing, and three-dimensional simulation are widely used. However, graphics processing units (GPUs) are used in high-speed operations and have limited general applications. In this study, we developed a task queue in a GPU to improve the performance of a web service using a multiprocessor and studied how to receive and process user requests in bulk. We propose the use of a GPGPU-based task queue to process user requests more than GPGPU based a central processing unit thread, and to process more GPU threads on task queue at about 136% to 233%, and proved that the proposed method is effective for web service.

EFFICIENT COMPUTATION OF COMPRESSIBLE FLOW BY HIGHER-ORDER METHOD ACCELERATED USING GPU (고차 정확도 수치기법의 GPU 계산을 통한 효율적인 압축성 유동 해석)

  • Chang, T.K.;Park, J.S.;Kim, C.
    • Journal of computational fluids engineering
    • /
    • v.19 no.3
    • /
    • pp.52-61
    • /
    • 2014
  • The present paper deals with the efficient computation of higher-order CFD methods for compressible flow using graphics processing units (GPU). The higher-order CFD methods, such as discontinuous Galerkin (DG) methods and correction procedure via reconstruction (CPR) methods, can realize arbitrary higher-order accuracy with compact stencil on unstructured mesh. However, they require much more computational costs compared to the widely used finite volume methods (FVM). Graphics processing unit, consisting of hundreds or thousands small cores, is apt to massive parallel computations of compressible flow based on the higher-order CFD methods and can reduce computational time greatly. Higher-order multi-dimensional limiting process (MLP) is applied for the robust control of numerical oscillations around shock discontinuity and implemented efficiently on GPU. The program is written and optimized in CUDA library offered from NVIDIA. The whole algorithms are implemented to guarantee accurate and efficient computations for parallel programming on shared-memory model of GPU. The extensive numerical experiments validates that the GPU successfully accelerates computing compressible flow using higher-order method.

GPU Accelerating Methods for Pease FFT Processing (Pease FFT 처리를 위한 GPU 가속 기법)

  • Oh, Se-Chang;Joo, Young-Bok;Kwon, Oh-Young;Huh, Kyung-Moo
    • Journal of Institute of Control, Robotics and Systems
    • /
    • v.20 no.1
    • /
    • pp.37-41
    • /
    • 2014
  • FFT (Fast Fourier Transform) has been widely used in various fields such as image processing, voice processing, physics, astronomy, applied mathematics and so forth. Much research has been conducted due to the importance of the FFT and recently new FFT algorithms using a GPU (Graphics Processing Unit) have been developed for the purpose of much faster processing. In this paper, the new optimal FFT algorithm using the Pease FFT algorithm has been proposed reflecting the hardware configuration of a GPGPU (General Purpose computing of GPU). According to the experiments, the proposed algorithm outperformed by between 3% to 43% compared to the CUFFT algorithm.

A Study of Depth Estimate using GPGPU in Monocular Image (GPGPU를 이용한 단일 영상에서의 깊이 추정에 관한 연구)

  • Yoo, Tae Hoon;Lee, Gang Seong;Park, Young Soo;Lee, Jong Yong;Lee, Sang Hun
    • Journal of Digital Convergence
    • /
    • v.11 no.12
    • /
    • pp.345-352
    • /
    • 2013
  • In this paper, a depth estimate method is proposed using GPU(Graphics Processing Unit) in monocular image. a monocular image is a 2D image with missing 3D depth information due to the camera projection and we used a monocular cue to recover the lost depth information by the projection present. The proposed algorithm uses an energy function which takes a variety of cues to create a more generalized and reliable depth map. But, a processing time is late because energy function is defined from the various monocular cues. Therefore, we propose a depth estimate method using GPGPU(General Purpose Graphics Processing Unit). The objective effectiveness of the algorithm is shown using PSNR(Peak Signal to Noise Ratio), a processing time is decrease by 61.22%.

New GPU computing algorithm for wind load uncertainty analysis on high-rise systems

  • Wei, Cui;Luca, Caracoglia
    • Wind and Structures
    • /
    • v.21 no.5
    • /
    • pp.461-487
    • /
    • 2015
  • In recent years, the Graphics Processing Unit (GPU) has become a competitive computing technology in comparison with the standard Central Processing Unit (CPU) technology due to reduced unit cost, energy and computing time. This paper describes the derivation and implementation of GPU-based algorithms for the analysis of wind loading uncertainty on high-rise systems, in line with the research field of probability-based wind engineering. The study begins by presenting an application of the GPU technology to basic linear algebra problems to demonstrate advantages and limitations. Subsequently, Monte-Carlo integration and synthetic generation of wind turbulence are examined. Finally, the GPU architecture is used for the dynamic analysis of three high-rise structural systems under uncertain wind loads. In the first example the fragility analysis of a single degree-of-freedom structure is illustrated. Since fragility analysis employs sampling-based Monte Carlo simulation, it is feasible to distribute the evaluation of different random parameters among different GPU threads and to compute the results in parallel. In the second case the fragility analysis is carried out on a continuum structure, i.e., a tall building, in which double integration is required to evaluate the generalized turbulent wind load and the dynamic response in the frequency domain. The third example examines the computation of the generalized coupled wind load and response on a tall building in both along-wind and cross-wind directions. It is concluded that the GPU can perform computational tasks on average 10 times faster than the CPU.

Research of accelerating method of video quality measurement program using GPGPU (GPGPU를 이용한 영상 품질 측정 프로그램의 가속화 연구)

  • Lee, Seonguk;Byeon, Gibeom;Kim, Kisu;Hong, Jiman
    • Smart Media Journal
    • /
    • v.5 no.4
    • /
    • pp.69-74
    • /
    • 2016
  • Recently, parallel computing using GPGPU(General-Purpose computing on Graphics Processing Units) according to the development of the graphics processing unit is expanding. This can be achieved through the processing speeds faster than traditional computing environments across many fields, including science, medicine, engineering, and analysis. However, in using the GPU technology to implement the a parallel program there are many constraints. In this paper, we port a CPU-based program(Video Quality Measurement Program) to use technology. The program ported to GPU-based show about 1.83 times the execution speed than CPU-based program. We study on the acceleration of the GPU-based program. Also we discuss the technical constraints and problems that occur when you modify the CPU to the GPU-based programs.

Implementation of $2{\times}2$ MIMO LTE Base Station using GPU for SDR System (GPU를 이용한 SDR 시스템 용 LTE MIMO 기지국 기능 구현)

  • Lee, Seung Hak;Kim, Kyung Hoon;Ahn, Chi Young;Choi, Seung Won
    • Journal of Korea Society of Digital Industry and Information Management
    • /
    • v.8 no.4
    • /
    • pp.91-98
    • /
    • 2012
  • This paper implements 2X2 MIMO Long Term Evolution (LTE) base station using Software defined radio (SDR) technology. The implemented base station system processes baseband signals on a Graphics Processor Unit(GPU). GPU is a high-speed parallel processor which provides very important advantage of using a very powerful C-based programming environment that is Compute Unified Device Architecture (CUDA). The implemented software-based base station system processes baseband signals through GPU. It utilizes USRP2 as its RF transceiver. In order to guarantee a real-time processing of LTE baseband signals, we have adopted well-known signal processing algorithms such as frame synchronization algorithms, ML detection, etc. using GPU operating in parallel processing.

Development and run time assessment of the GPU accelerated technique of a 2-Dimensional model for high resolution flood simulation in wide area (광역 고해상도 홍수모의를 위한 2차원 모형의 GPU 가속기법 개발 및 실행시간 평가)

  • Choi, Yun Seok;Noh, Hui Seong;Choi, Cheon Kyu
    • Journal of Korea Water Resources Association
    • /
    • v.55 no.12
    • /
    • pp.991-998
    • /
    • 2022
  • The purpose of this study is to develop GPU (Graphics Processing Unit) acceleration technique for 2-dimensional model and to assess the effectiveness for high resolution flood simulation in wide area In this study, GPU acceleration technique was implemented in the G2D (Grid based 2-Dimensional land surface flood model) model, using implicit scheme and uniform square grid, by using CUDA. The technique was applied to flood simulation in Jinju-si. The spatial resolution of the simulation domain is 10 m × 10 m, and the number of cells to calculate is 5,090,611. Flood period by typhoon Mitag, December 2019, was simulated. Rainfall radar data was applied to source term and measured discharge of Namgang-Dam (Ilryu-moon) and measured stream flow of Jinju-si (Oksan-gyo) were applied to boundary conditions. From this study, 2-dimensional flood model could be implemented to reproduce the measured water level in Nam-gang (Riv.). The results of GPU acceleration technique showed more faster flood simulation than the serial and parallel simulation using CPU (Central Processing Unit). This study can contribute to the study of developing GPU acceleration technique for 2-dimensional flood model using implicit scheme and simulating land surface flood in wide area.

Design Conditions for Parallel Sorting Algorithms using GPU (GPU를 사용한 병렬 정렬 알고리즘의 설계 조건)

  • Lee, Ho-Suk
    • Proceedings of the Korean Information Science Society Conference
    • /
    • 2011.06b
    • /
    • pp.1-4
    • /
    • 2011
  • 본 논문에서는 병렬 정렬(sorting) 알고리즘들에 대하여 논의한다. 정렬 알고리즘은 컴퓨터 과학에서 매우 중요한 위치를 차지하며 지난 50~60년 동안 많은 연구자들에 의하여 연구되었다. 10년 전에는 GPU(Graphics Processing Unit) 병렬 프로세서가 개발되어 병렬 정렬 알고리즘에 대한 연구도 활발히 진행되고 있다. 병렬 정렬 알고리즘은 대체적으로 bitonic 정렬, radix 정렬, merge 정렬, 혹은 이들 정렬 알고리즘들을 혼합하여 사용한 방법으로 분류된다. 논문에서는 GPU를 사용한 새로운 효율적인 병렬 정렬 알고리즘의 설계 조건을 논의한다.

A Study on Efficiency of Cryptography Used by CPU and GPU (CPU와 GPU를 이용한 암호화 효율성 연구)

  • Byeon, Jin-Yeong;Lee, Ki-Young
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2012.05a
    • /
    • pp.678-680
    • /
    • 2012
  • 1970년대 라디오 주파수를 사용하여 컴퓨터 통신 네트워크가 구축된 이후 눈부신 발전을 거듭하여 Personal Computer 뿐만 아니라 Mobile이나 Tablet PC등에서도 인터넷이 가능하다. 이렇게 다양한 매체를 통해 인터넷을 사용함에 따라 보안에 대한 중요성이 높아지고 있다. 하지만 최근 현대 캐피탈이나 농협, 네이트와 같은 해킹 사례를 보면 평문 데이터 사용에 의해 피해가 더욱 확대 되었다. 평문 데이터 사용함에 따라 보안 위협이 커지는데 평문 데이터를 사용하는 이유를 암호화를 사용했을 때보다 QoS 하락 때문이라고 볼 수있다. 이를 해결하기 위해 고정된 인프라에서 잉여 자원인 GPU를 사용하여 암호화를 할 때 QoS 하락을 줄일 수 있을 것이다. 또한 CPU보다는 멀티코어를 사용한 병렬 처리를 활용하여 CPU보다 상대적으로 효율적인 암호화가 가능하다고 생각한다. 본 논문에서는 CPU를 이용한 암호화 처리 속도와 GPU를 이용한 암호화 처리 속도를 비교하여 GPU를 이용한 암호화 처리 가능성을 검토하였다.

  • PDF