• Title/Summary/Keyword: PCI-Express (PCIe)

Search Result 8, Processing Time 0.023 seconds

Implementation and Performance Evaluation of PCI express on Xilinx FPGA (Xilinx FPGA용 PCI express 구현 및 성능 분석)

  • Lee, Jin
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.22 no.12
    • /
    • pp.1667-1674
    • /
    • 2018
  • Recently, speeding up real time calculation using the specialized hardware accelerator is often used in the various engineering and science area, and the accelerators are required to include PCI express interconnection between FPGA and a host computer. The implementation of the high speed PCIe for the multi-giga bytes per second transmission is one of the most difficult issue in the development of the accelerators. There are several commercialized IP solutions and research results in the literature, but these solutions are required extra cost and design period to analyze the detailed implementation method. For the hardware accelerator on Xilinx FPGA, utilizing Xilinx's XDMA PCIe IP, which is provided without extra charge, can be the best solution in terms of the development period and cost. Consequently, this paper presents the evaluation system on Zynq-7000 FPGA and Windows 10 host computer, and analyze the performance of the PCIe IP with various configuration parameters.

Design and Implementation of an Alternate System Interconnect based on PCI Express (PCI Express 기반 시스템 인터커넥트의 설계 및 구현)

  • Kim, Young Woo;Ren, Ye;Choi, WonHyuk
    • Journal of the Institute of Electronics and Information Engineers
    • /
    • v.52 no.8
    • /
    • pp.74-85
    • /
    • 2015
  • PCI Express is a well-known and widely used de-facto system bus standard for connecting among a processor and IO devices. PCI Express is originated from old PCI standard, and its most of applications are limited to be used within a PC or server system. But, because of its fast speed, low power consumption, and good protocol efficiency, it is considered as one of a good candidate for an alternate system interconnect for many years. In this paper, we present design, implementation and early evaluation of an alternate system interconnect by utilizing PCI Express. The developed alternate system interconnect using PCI Express (named PCIeLINK) utilizes non-transparent bridging (NTB) technic which generally used in fail-over system in PCI and PCI Express. By using NTB technic, PCI Express device can be extended to outside of a system without electrical and logical problems arising during system boot and enumeration. To build up an alternate system interconnect, we designed and implemented a network interface card having multiple PCI Express ${\times}4$ connections (theoretically 20 Gbps) and tested, The early test results revealed that an ${\times}4$ port in the card showed 8.6 Gbps peak performance for bulk transmission and 5.1 Gbps peak for normal TCP/IP transfer.

Implementation of Multipurpose PCI Express Adapter Cards with On-Board Optical Module

  • Koo, Kyungmo;Yu, Junglok;Kim, Sangwan;Choi, Min;Cha, Kwangho
    • Journal of Information Processing Systems
    • /
    • v.14 no.1
    • /
    • pp.270-279
    • /
    • 2018
  • PCI Express (PCIe) bus, which was only used as an internal I/O bus of a computer system, has expanded its function to outside of a system, with progress of PCIe switching processor. In particular, advanced features of PCIe switching processor enable PCIe bus to serve as an interconnection network as well as connecting external devices. As PCIe switching processors more advanced, it is required to consider the different adapter card architecture. This study developed multipurpose adapter cards by applying an on-board optical module, a latest optical communications element, in order to improve transfer distance and utilization. The performance evaluation confirmed that the new adapter cards with long cable can provide the same bandwidth as that of the existing adapter cards with short copper cable.

Design and Implementation of OpenSHMEM-Light using PCIe NTB (PCIe NTB를 활용한 OpenSHMEM-Light의 설계 및 구현)

  • Ju, Youngwoong;Choi, Min
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2016.10a
    • /
    • pp.58-61
    • /
    • 2016
  • .PCI Express는 고속, 저전력 등의 특성으로 업계 표준으로서 많이 쓰이고 있는 프로세서와 주변 I/O 장치들을 연결하는 버스 기술이다. 또한, PCI Express는 인피니밴드와 이더넷과 더불어 고성능 컴퓨터나 컴퓨터 클러스터를 위한 시스템 인터커넥트 기술로 널리 쓰이고 있다. PGAS(partitioned global address space) 프로그래밍 모델은 컴퓨터 클러스터와 같은 다중 호스트 시스템에서 단측 RDMA(remote direct memory access)를 구현하는데 많이 이용된다. 본 논문에서는 PCI Express 기반 RDMA를 구현하기 위해 PGAS 프로그래밍 모델인 OpenSHMEM의 기존의 특징을 유지하여 PCI Express 기반 OpenSHMEM API를 설계 및 구현하였다. 구현한 OpenSHMEM API는 PCI Express의 NTB(non-transparent bridge) 기술로 2대의 PC를 연결한 시스템에서 매트릭스 곱셈 예제를 통하여 실험하였다.

A Study on PCIe Switching Technology for Inter-node Communications (노드간 통신을 위한 PCIe 스위칭 기술 연구)

  • CHA, Kwangho;Yu, Junglok;Kim, Sungho
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2015.10a
    • /
    • pp.56-58
    • /
    • 2015
  • 버스 시스템은 단일 시스템 내부의 디바이스들을 물리적으로 연결하는데 사용되던 연결망으로 최근에는 PCI-Express(PCIe)방식의 버스가 주로 사용되고 있다. PCIe 스위칭 기술은 단순히 노드내의 디바이스들만을 연결하는데 그치지 않고 더 나아가 노드간 통신을 가능하게 하는 기능을 제공한다. 본 연구에서는 PCIe 스위칭 기술의 특징과 노드간 통신에 사용될 수 있는 다양한 PCIe 스위치 활용 기법을 살펴보았다.

Application-level Performance Evaluation of Transmission Media for PCIe Expansion (PCIe 확장 보드용 통신 매체에 대한 응용 레벨 성능 평가)

  • CHA, Kwangho;Koo, Kyungmo;Jung, Hyun Mi
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2019.10a
    • /
    • pp.2-4
    • /
    • 2019
  • PCI Express(PCIe) 버스는 시스템 내·외부의 다양한 디바이스들을 연결하는데 여전히 중요한 역할을 하고 있다. 이 PCIe 버스를 확장하기 위해서 PCIe 스위치가 사용되며 호스트(서버)와 외부 디바이스를 연결할 때 다양한 케이블 미디어가 사용된다. 본 연구에서는 광 타입과 구리선 타입의 2종류 케이블이 장착 가능한 자체 제작된 PCIe 어댑터 카드를 사용하여 응용 프로그램 레벨에서 성능을 확인하였다. 서로 다른 2종류의 디바이스들을 대상으로 실험한 결과, 성능상의 큰 차이는 발견되지 않아서 케이블의 종류와는 무관하게 상호 운영에 크게 문제가 없을 것으로 예상된다.

Analysis of I/O Response Time Throughout NVMe Driver Implementation Architectures (NVMe 드라이버 구현 방식에 따른 I/O 응답시간 분석)

  • Kang, Ingu;Joo, Yongsoo;Lim, Sung-Soo
    • IEMEK Journal of Embedded Systems and Applications
    • /
    • v.12 no.3
    • /
    • pp.139-147
    • /
    • 2017
  • In recent years, non-volatile memory express (NVMe), a new host controller interface standard, has been adapted to overcome performance bottlenecks caused by the acceleration of solid state drives (SSD). Recently, performance breakthrough cases over AHCI based SATA SSDs by adapting NVMe based PCI Express (PCIe) SSD to servers and PCs have been reported. Furthermore, replacing legacy eMMC-flash storage with NVMe based storage is also considered for next generation of mobile devices such as smartphones. The Linux kernel includes drivers for NVMe support, and as the kernel version increases, the implementation of the NVMe driver code has changed. However, mobile devices are often equipped with older versions of Android operating systems (OSes), where the newest features of NVMe drivers are not available. Therefore, different features of different NVMe driver implementations are not well evaluated on Android OSes. In this paper, we analyze the response time of the NVMe driver for various Linux kernel version.

Analyzing the Impact of Supply Noise on Jitter in GBPS Serial Links on a Merged I/O-Core Power Delivery Network

  • Tan, Fern-Nee;Lee, Sheng Chyan
    • Journal of the Microelectronics and Packaging Society
    • /
    • v.20 no.4
    • /
    • pp.69-74
    • /
    • 2013
  • In this paper, the impact of integrating large number of I/O (Input-Output) and Core power Delivery Network (PDN) on a 6 layers Flip-Chip Ball Grid Array (FCBGA) package is investigated. The impact of core induced supply noise on high-speed I/O interfaces, and high-speed I/O interface's supply noise coupling to adjacent high-speed I/O interfaces' jitter impact are studied. Concurrent stress validation software is used to induce SSO noise on each individual I/O interfaces; and at the same time; periodic noise is introduced from Core PDN into the I/O PDN domain. In order to have the maximum coupling impact, a prototype package is designed to merge the I/O and Core PDN as one while impact on jitter on each I/O interfaces are investigated. In order to understand the impact of the Core to I/O and I/O to I/O noise, the on-die noise measurements were measured and results were compared with the original PDN where each I/O and Core PDN are standalone and isolated are used as a benchmark.