• Title/Summary/Keyword: 인피니밴드

Search Result 22, Processing Time 0.035 seconds

All Flash Array Storage Virtualisation using SCST (SCST를 이용한 All Flash Array 스토리지 가상화)

  • Heo, Huiseong;Pirahandeh, Mehdi;Lee, Kwangsoo;Kim, Deokhwan
    • KIISE Transactions on Computing Practices
    • /
    • v.20 no.10
    • /
    • pp.525-533
    • /
    • 2014
  • SCST(The generic SCSI target subsystem for Linux) enables developers to make SCSI target storage and supports various SCSI network protocol such as iSCSI, FC, SRP. In this paper, we propose storage virtualization method using SCST and virtualize all flash array as high performance storage through 4Gb Fiber Channel, 10Gb Ethernet and 40Gb Infiniband and evaluate their performance, respectively. Experimental result shows that 40Gb infiniband network appliance have better performance than others. In case of sequential/random read, 40Gb infiniband network appliance shows 78% and 79% of local all flash array performance attached to SCSI target system. In case of sequential/random write, it shows 83% and 88% of local flash array performance attached to SCSI target system.

The Technology Trend of Interconnection Network for High Performance Computing (고성능 컴퓨팅을 위한 인터커넥션 네트워크 기술 동향)

  • Cho, Hyeyoung;Jun, Tae Joon;Han, Jiyong
    • Journal of the Korea Convergence Society
    • /
    • v.8 no.8
    • /
    • pp.9-15
    • /
    • 2017
  • With the development of semiconductor integration technology, central processing units and storage devices have been miniaturized and performance has been rapidly developed, interconnection network technology is becoming a more important factor in terms of the performance of high performance computing system. In this paper, we analyze the trend of interconnection network technology used in high performance computing. Interconnect technology, which is the most widely used in the Supercomputer Top 500(2017. 06.), is an Infiniband. Recently, Ethernet is the second highest share after InfiniBand due to the emergence of 40/100Gbps Gigabit Ethernet technology. Gigabit Ethernet, where latency performance is lower than InfiniBand, is preferred in cost-effective medium-sized data centers. In addition, top-end HPC systems that demand high performance are devoting themselves from Ethernet and InfiniBand technologies and are attempting to maximize system performance by introducing their own interconnect networks. In the future, high-performance interconnects are expected to utilize silicon-based optical communication technology to exchange data with light.

PERFORMANCE ANALYSIS OF THE PARALLEL CUPID CODE IN DISTRIBUTED MEMORY SYSTEM BASED ETHERNET AND INFINIBAND NETWORK (이더넷과 인피니밴드 네트워크 기반의 분산 메모리 시스템에서 병렬성능 분석)

  • Jeon, B.J.;Choi, H.G.
    • Journal of computational fluids engineering
    • /
    • v.19 no.2
    • /
    • pp.24-29
    • /
    • 2014
  • In this study, a parallel performance of CUPID-code has been investigated for both Ethernet and Infiniband network system to examine the effect of cache memory and network-speed. Bi-conjugate gradient solver of CUPID-code has been parallelised by using domain decomposition method and message passing interface (MPI). It is shown that the parallel performance of Ethernet-network system is worse than that of Infiniband-network system due to the slow network-speed and a small cache memory. It is also found that the parallel performance of each system deteriorates for a small problem due to the communication overhead, but the performance of Infiniband-network system is better than Ethernet-network system due to a much faster network-speed. For a large problem, the parallel performance depends less on network system.

Tactical and Operational Control Platform based on OpenStack Cloud and Data Distribution Service (오픈스택 클라우드 및 데이터 분산 서비스 기반 전술작전통제 플랫폼)

  • Jun, Tae Joon;Yoo, Myong Hwan;Cho, Kyu Tae;Lee, Seung Young;Yeun, Kyuoke;Kim, Daeyoung
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2015.10a
    • /
    • pp.199-201
    • /
    • 2015
  • 국방 분야에서 전술작전센터를 구성하는 전술작전통제 플랫폼은 다양한 내/외부 체계 시험 장치를 연동하여 교전/작전 통제를 수행하는 플랫폼이다. 기존의 전술작전통제 플랫폼은 서로 다른 환경에서 개발된 시험 장치 소프트웨어를 통합 구축하는데 어려움이 있으며, 플랫폼을 구성하는 서버 자원의 탄력적 분배가 불가능하다. 본 연구는 전술작전센터를 구성하는 전술작전통제 플랫폼을 오픈스택 클라우드 및 KVM 하이퍼바이저를 기반으로 구축하여 내/외부 체계 시험 장치 소프트웨어의 실행 환경을 가상 머신으로 제공하였으며, 가상 머신 간 고속 데이터 통신을 위해 데이터 분산 서비스 및 10Gbps 이더넷, QDR 인피니밴드 네트워크를 활용하였다. 또한 오픈스택 클라우드 기반 전술작전통제 플랫폼의 네트워크를 구성하는 10Gbps 이더넷 및 QDR 인피니밴드 네트워크상에서 데이터 분산 서비스의 성능 평가를 수행하였다.

Application level performance evaluation of Infiniband Network (인피니밴드 네트워크에 대한 응용 레벨 성능 분석)

  • Cha, Kwang-Ho;Kim, Sung-Ho
    • Proceedings of the Korean Information Science Society Conference
    • /
    • 2005.11a
    • /
    • pp.1003-1005
    • /
    • 2005
  • 클러스터 시스템의 성능과 관련되어 중요성이 강조되는 분야로 SAN(System Area Network)을 이야기할 수 있다. 특히 infiniband 제품의 출시는 이 분야에 대한 관심을 더욱 집중시키는 계기가 되었다. 이에 본 논문에서는 Infiniband가 보여주는 단순한 네트워크적인 특징 이외에 응용 프로그램의 성능에 어떠한 영향을 미치는가에 대하여 실험하고 분석하였다.

  • PDF

Implementation of Storage Service Protocol on Infiniband based Network (인피니밴드 네트웍에서 RDMA 기반의 저장장치 서비스 프로토콜개발)

  • Joen Ki-Man;Park Chang-Won;Kim Young-Hwan
    • 한국정보통신설비학회:학술대회논문집
    • /
    • 2006.08a
    • /
    • pp.77-81
    • /
    • 2006
  • Because of the rapid increasing of network user, there are some problems to tolerate the network overhead. Recently, the research and technology of the user-level for high performance and low latency than TCP/IP which relied upon the kernel for processing the messages. For example, there is an Infiniband technology. The Infiniband Trade Association (IBTA) has been proposed as an industry standard for both communication between processing node and I/O devices and for inter-processor communication. It replaces the traditional bus-based interconnect with a switch-based network for connecting processing node and I/O devices. Also Infiniband uses RDMA (Remote DMA) for low latency of CPU and OS to communicate between Remote nodes. In this paper, we develop the SRP (SCSI RDMA Protocol) which is Storage Access Protocol on Infiniband network. And will compare to FC (Fibre Channle) based I-SCSI (Internet SCSI) that it is used to access storage on Etherent Fabric.

  • PDF

Design and Implementation of OpenSHMEM-Light using PCIe NTB (PCIe NTB를 활용한 OpenSHMEM-Light의 설계 및 구현)

  • Ju, Youngwoong;Choi, Min
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2016.10a
    • /
    • pp.58-61
    • /
    • 2016
  • .PCI Express는 고속, 저전력 등의 특성으로 업계 표준으로서 많이 쓰이고 있는 프로세서와 주변 I/O 장치들을 연결하는 버스 기술이다. 또한, PCI Express는 인피니밴드와 이더넷과 더불어 고성능 컴퓨터나 컴퓨터 클러스터를 위한 시스템 인터커넥트 기술로 널리 쓰이고 있다. PGAS(partitioned global address space) 프로그래밍 모델은 컴퓨터 클러스터와 같은 다중 호스트 시스템에서 단측 RDMA(remote direct memory access)를 구현하는데 많이 이용된다. 본 논문에서는 PCI Express 기반 RDMA를 구현하기 위해 PGAS 프로그래밍 모델인 OpenSHMEM의 기존의 특징을 유지하여 PCI Express 기반 OpenSHMEM API를 설계 및 구현하였다. 구현한 OpenSHMEM API는 PCI Express의 NTB(non-transparent bridge) 기술로 2대의 PC를 연결한 시스템에서 매트릭스 곱셈 예제를 통하여 실험하였다.

Simulation of Subnet Management for InfiniBand (채널 기반 인피니밴드의 서브넷 관리를 위한 시뮬레이션)

  • Kim, Young-Hwan;Youn, Hee-Yong;Park, Chang-Won;Lee, Hyoung-Su;Go, Jae-Jin;Park, Sang-Hyun
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2002.11a
    • /
    • pp.535-538
    • /
    • 2002
  • InfiniBand is a switched-fabric architecture for next generation I/O systems and data centers. The InfiniBand Architecture (IBA) promises to replace bus-based architectures, such as PCI, with a switched-based fabric whose benefits include higher performance, higher RAS (reliability, availability, scalability), and the ability to create modular networks of servers and shared I/O devices. The switched-fabric InfiniBand consists of InfiniBand subnets with channel adapters, switches, and routers. In order to fully grasp the operational characteristics of InfiniBand architecture (IBA) and use them in ongoing design specification, simulation of subnet management of IBA is inevitable. In this paper, thus, we implement an IBA simulator and test some practical sample networks using it. The simulator shows the flow of operation by which the correctness and effectiveness of the system can be verified.

  • PDF

Large-Memory Data Processing on a Remote Memory System using Commodity Hardware (대용량 메모리 데이타 처리를 위한 범용 하드웨어 기반의 원격 메모리 시스템)

  • Jung, Hyung-Soo;Han, Hyuck;Yeom, Heon-Y.
    • Journal of KIISE:Computer Systems and Theory
    • /
    • v.34 no.9
    • /
    • pp.445-458
    • /
    • 2007
  • This article presents a novel infrastructure for large-memory database processing using commodity hardware with operating system support. We exploit inexpensive PCs and a high-speed network capable of Remote Direct Memory Access (RDMA) operations to build a new memory hierarchy between fast volatile memory and slow disk storage. The new memory hierarchy guarantees a reasonable response time, and its storage size enables us to run large-memory database systems with little performance degradation. The proposed architecture has two main components: (1) a remote memory system inside the Linux kernel to manage other computers' memory pages efficiently and (2) a remote memory pager responsible for manipulating remote read/write operations on remote memory pages. We insist that the proposed architecture is practical enough to support the rigorous demands of commercial in-memory database systems by demonstrating the performance of publicly available main-memory databases (e.g., MySQL) on our prototyped system. The experimental results show very interesting results from the TPC-C benchmark.