• Title/Summary/Keyword: Access Latency

Search Result 330, Processing Time 0.023 seconds

Random Number Generation using SDRAM (SDRAM을 사용한 난수 발생)

  • Pyo, Chang-Woo
    • Journal of KIISE:Computing Practices and Letters
    • /
    • v.16 no.4
    • /
    • pp.415-420
    • /
    • 2010
  • Cryptographic keys for security should be generated by true random number generators that apply irreversible hashing algorithms to initial values taken from a random source. As DRAM shows randomness in its access latency, it can be used as a random source. However, systems with synchronous DRAM (SDRAM) do not easily expose such randomness resulting in highly clustered random numbers. We resolved this problem by using the xor instruction. Statistical testing shows that the generated random bits have the quality comparable to true random bit sequences. The performance of bit generation is at the order of 100 Kbits/sec. Since the proposed random number generation requires neither external devices nor any special circuits, this method may be used in any computing device that employs DRAM.

Ultra-low-latency services in 5G systems: A perspective from 3GPP standards

  • Jun, Sunmi;Kang, Yoohwa;Kim, Jaeho;Kim, Changki
    • ETRI Journal
    • /
    • v.42 no.5
    • /
    • pp.721-733
    • /
    • 2020
  • Recently, there is an increasing demand for ultra-low-latency (ULL) services such as factory automation, autonomous driving, and telesurgery that must meet an end-to-end latency of less than 10 ms. Fifth-generation (5G) New Radio guarantees 0.5 ms one-way latency, so the feasibility of ULL services is higher than in previous mobile communications. However, this feasibility ensures performance at the radio access network level and requires an innovative 5G network architecture for end-to-end ULL across the entire 5G system. Hence, we survey in detailed two the 3rd Generation Partnership Party (3GPP) standardization activities to ensure low latency at network level. 3GPP standardizes mobile edge computing (MEC), a low-latency solution at the edge network, in Release 15/16 and is standardizing time-sensitive communication in Release 16/17 for interworking 5G systems and IEEE 802.1 time-sensitive networking (TSN), a next-generation industry technology for ensuring low/deterministic latency. We developed a 5G system based on 3GPP Release 15 to support MEC with a potential sub-10 ms end-to-end latency in the edge network. In the near future, to provide ULL services in the external network of a 5G system, we suggest a 5G-IEEE TSN interworking system based on 3GPP Release 16/17 that meets an end-to-end latency of 2 ms.

A Locality-Aware Write Filter Cache for Energy Reduction of STTRAM-Based L1 Data Cache

  • Kong, Joonho
    • JSTS:Journal of Semiconductor Technology and Science
    • /
    • v.16 no.1
    • /
    • pp.80-90
    • /
    • 2016
  • Thanks to superior leakage energy efficiency compared to SRAM cells, STTRAM cells are considered as a promising alternative for a memory element in on-chip caches. However, the main disadvantage of STTRAM cells is high write energy and latency. In this paper, we propose a low-cost write filter (WF) cache which resides between the load/store queue and STTRAM-based L1 data cache. To maximize efficiency of the WF cache, the line allocation and access policies are optimized for reducing energy consumption of STTRAM-based L1 data cache. By efficiently filtering the write operations in the STTRAM-based L1 data cache, our proposed WF cache reduces energy consumption of the STTRAM-based L1 data cache by up to 43.0% compared to the case without the WF cache. In addition, thanks to the fast hit latency of the WF cache, it slightly improves performance by 0.2%.

Implementation of Storage Service Protocol on Infiniband based Network (인피니밴드 네트웍에서 RDMA 기반의 저장장치 서비스 프로토콜개발)

  • Joen Ki-Man;Park Chang-Won;Kim Young-Hwan
    • 한국정보통신설비학회:학술대회논문집
    • /
    • 2006.08a
    • /
    • pp.77-81
    • /
    • 2006
  • Because of the rapid increasing of network user, there are some problems to tolerate the network overhead. Recently, the research and technology of the user-level for high performance and low latency than TCP/IP which relied upon the kernel for processing the messages. For example, there is an Infiniband technology. The Infiniband Trade Association (IBTA) has been proposed as an industry standard for both communication between processing node and I/O devices and for inter-processor communication. It replaces the traditional bus-based interconnect with a switch-based network for connecting processing node and I/O devices. Also Infiniband uses RDMA (Remote DMA) for low latency of CPU and OS to communicate between Remote nodes. In this paper, we develop the SRP (SCSI RDMA Protocol) which is Storage Access Protocol on Infiniband network. And will compare to FC (Fibre Channle) based I-SCSI (Internet SCSI) that it is used to access storage on Etherent Fabric.

  • PDF

MBS-LVM: A High-Performance Logical Volume Manager for Memory Bus-Connected Storages over NUMA Servers

  • Lee, Yongseob;Park, Sungyong
    • Journal of Information Processing Systems
    • /
    • v.15 no.1
    • /
    • pp.151-158
    • /
    • 2019
  • With the recent advances of memory technologies, high-performance non-volatile memories such as non-volatile dual in-line memory module (NVDIMM) have begun to be used as an addition or an alternative to server-side storages. When these memory bus-connected storages (MBSs) are installed over non-uniform memory access (NUMA) servers, the distance between NUMA nodes and MBSs is one of the crucial factors that influence file processing performance, because the access latency of a NUMA system varies depending on its distance from the NUMA nodes. This paper presents the design and implementation of a high-performance logical volume manager for MBSs, called MBS-LVM, when multiple MBSs are scattered over a NUMA server. The MBS-LVM consolidates the address space of each MBS into a single global address space and dynamically utilizes storage spaces such that each thread can access an MBS with the lowest latency possible. We implemented the MBS-LVM in the Linux kernel and evaluated its performance by porting it over the tmpfs, a memory-based file system widely used in Linux. The results of the benchmarking show that the write performance of the tmpfs using MBS-LVM has been improved by up to twenty times against the original tmpfs over a NUMA server with four nodes.

Cache Sensitive T-tree Index Structure (캐시를 고려한 T-트리 인덱스 구조)

  • Lee Ig-hoon;Kim Hyun Chul;Hur Jae Yung;Lee Snag-goo;Shim JunHo;Chang Juho
    • Journal of KIISE:Databases
    • /
    • v.32 no.1
    • /
    • pp.12-23
    • /
    • 2005
  • In the past decade, advances in speed of commodity CPUs have iu out-paced advances in memory latency Main-memory access is therefore increasingly a performance bottleneck for many computer applications, including database systems. To reduce memory access latency, cache memory incorporated in the memory subsystem. but cache memories can reduce the memory latency only when the requested data is found in the cache. This mainly depends on the memory access pattern of the application. At this point, previous research has shown that B+ trees perform much faster than T-trees because B+ trees are more cache conscious than T-trees, and also proposed 'Cache Sensitive B+trees' (CSB. trees) that are more cache conscious than B+trees. The goal of this paper is to make T-trees be cache conscious as CSB-trees. We propose a new index structure called a 'Cache Sensitive T-trees (CST-trees)'. We implemented CST-trees and compared performance of CST-trees with performance of other index structures.

Perfomance Evaluation of efficent handover Latency Using MIH Services in MIPv4 (MIH를 이용한 효율적인 MIPv4망의 구성에 관한 연구)

  • Kim, Ki-Yong;Jang, Jong-Wook
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2007.10a
    • /
    • pp.75-78
    • /
    • 2007
  • Mobile IP provides hand-held devices with mobility which allows the user to do work over the network. However, handover time due transfer between access routers causes network delays and data loss. L2Trigger Handover expects this handover to take place, and executes L3 handover before L2 handover takes place, thereby reducing overall handover latency, although it still is an issue since handover latency between AR is not completely eliminated in L2 trigger handover. In this paper took into consideration where MIH is used in MIPv4 and using MIH Table when handover is about to occur in MN(Mobile Node), thereby pre-fetching data needed by Handover. In this way, when the handover is estimated, it improves the init time that L2trigger had. Furthermore we can find that we can execute the handover with shorten init time in smaller and narrow overlap length

  • PDF

Implementation of Light-weight I/O Stack for NVMe-over-Fabrics

  • Ahn, Sungyong
    • International journal of advanced smart convergence
    • /
    • v.9 no.3
    • /
    • pp.253-259
    • /
    • 2020
  • Most of today's large-scale cloud systems and enterprise data centers are distributing resources to improve scalability and resource utilization. NVMe-over-Fabric protocol allows submitting NVMe commands to a remote NVMe SSD through RDMA (Remote Direct Memory Access) network. It is attracting attention recently because it is possible to construct a disaggregation storage system with low latency through the protocol. However, the current I/O stack of NVMe-over-Fabric has an inefficient structure for maintaining compatibility with the traditional I/O stack. Therefore, in this paper, we propose a new mechanism to reduce I/O latency and CPU overhead by modifying I/O path of NVMe-over-Fabric to pass through legacy block layer. According to the performance evaluation results, the proposed mechanism is able to reduce the I/O latency and CPU overhead by up to 22% and 24% compared to the existing NVMe-over-Fabrics protocol, respectively.

Low Handover Latency for WiBro Network

  • Tae Ryoo-Kyoo;Park Se-Jun;Roh Jae-Hoon
    • 한국정보통신설비학회:학술대회논문집
    • /
    • 2006.08a
    • /
    • pp.43-46
    • /
    • 2006
  • IEEE 802.16 WirelessMAN aiming to broadband wireless access (BWA) is evolving to 4G mobile communication system through the standardization of IEEE 802.16e supporting mobility on existing fixed WirelessMAN system. It is necessary for hand-over to provide seamless data service while MS (Mobile Station) moves to another BS (Base Station). Because the performance of handover affects packet loss or delay of any communications, it must consider low latency handover mechanism in packet based network. In this paper, we describes handover scheme of IEEE 802.16e with the cell edge interference problem and shows the way to solve the problem in frequency reuse one deployment. Our scheme reduces the handover latency and packet loss probability.

  • PDF

Efficient Message Scheduling for WDM Optical Networks with Minimizing Flow Time

  • Huang, Xiaohong;Ma, Maode
    • Journal of Communications and Networks
    • /
    • v.6 no.2
    • /
    • pp.147-155
    • /
    • 2004
  • In this paper, we propose an efficient sequencing technique, namely minimum Row time scheduling (MFTS), to manage variable-Iength message transmissions for single-hop passive starcoupled WDM optical networks. By considering not only the message length but also the state of the receivers and the tuning latency, the proposed protocol can reduce the average delay of the network greatly. This paper also introduces a new channel assignment technique latency minimizing scheduling (LMS), which aims to reduce the scheduling latency. We evaluate the proposed algorithm, using extensive discrete-event simulations, by comparing its performance with shortest job first (SJF) algorithm. We find that significant improvement in average delay could be achieved by MFTS algorithm. By combining the proposed message sequencing technique with the channel selection technique, the performance of the optical network could be further improved.