• Title/Summary/Keyword: I/O latency

Search Result 75, Processing Time 0.024 seconds

I/O Translation Layer Technology for High-performance and Compatibility Using New Memory (뉴메모리를 이용한 고성능 및 호환성을 위한 I/O 변환 계층 기술)

  • Song, Hyunsub;Moon, Young Je;Noh, Sam H.
    • Journal of KIISE
    • /
    • v.42 no.4
    • /
    • pp.427-433
    • /
    • 2015
  • The rapid advancement of computing technology has triggered the need for fast data I/O processing and high-performance storage technology. Next generation memory technology, which we refer to as new memory, is anticipated to be used for high-performance storage as they have excellent characteristics as a storage device with non-volatility and latency close to DRAM. This research proposes NTL (New memory Translation layer) as a technology to make use of new memory as storage. With the addition of NTL, conventional I/O is served with existing mature disk-based file systems providing compatibility, while new memory I/O is serviced through the NTL to take advantage of the byte-addressability feature of new memory. In this paper, we describe the design of NTL and provide experiment measurement results that show that our design will bring performance benefits.

RFJ: A Reliable and Fast Journaling Mechanism (RFJ: 신뢰적 고성능 데이터 버퍼 저널링 기법)

  • Park, Sejin
    • Journal of the Korea Academia-Industrial cooperation Society
    • /
    • v.20 no.7
    • /
    • pp.45-51
    • /
    • 2019
  • Modern file systems have journaling mechanism to maintain their stored state consistently even under unexpected system crashes or disasters. However, the journaling makes I/O throughput lower. This performance degradation comes from the ordering mechanism between the data buffer and metadata buffer and two-staged buffer writing. Especially, if the data buffer and metadata buffer are journalled at the same time, then it incurs significant performance degradation due to the two-staged writing. That shows the trade-off relation-ship between I/O performance and system reliability. In this paper, we propose RFJ: a reliable and fast jour-naling mechanism to deal with this trade-off relationship. We propose an ordering enforced writeback journaling mode and selective journaling mechanism. The Ordering enforced writeback journaling mode achieves low I/O latency and the selective journaling mechanism achieves high reliability. The experimental result shows that the performance of RFJ is almost 5x faster than the journal mode of Ext3 file system but it still supports the same reliability with the journal mode.

Register-Based Parallel Pipelined Scheme for Synchronous DRAM (동기식 기억소자를 위한 레지스터를 이용한 병렬 파이프라인 방식)

  • Song, Ho Jun
    • Journal of the Korean Institute of Telematics and Electronics A
    • /
    • v.32A no.12
    • /
    • pp.108-114
    • /
    • 1995
  • Recently, along wtih the advance of high-performance system, synchronous DRAM's (SDRAM's) which provide consecutive data output synchronized with an external clock signal, have been reported. However, in the conventional SDRAM's which utilize a multi-stage serial pipelined scheme, the column path is divided into multi-stages depending on CAS latency N. Thus, as the operating speed and CAS latency increase, new stages must be added, thereby causing a large area penalty due to additinal latches and I/O lines. In the proposed register-based parallel pipelined scheme, (N-1) registers are located between the read data bus line pair and the data output buffer and the coming data are sequentially stored. Since the column data path is not divided and the read data is directly transmitted to the registers, the busrt read operation can easily be achieved at higher frequencies without a large area penalty and degradation of internal timing margin. Simulation results for 0.32um-Tech. 4-Bank 64M SDRAM show good operation at 200MHz and an area increment is less than 0.1% when CAS latency N is increased from 3 to 4.. This pipelined scheme is more advantageous as the operating frequency increases.

  • PDF

Demand-based FTL Cache Partitioning for Large Capacity SSDs (대용량 SSD를 위한 요구 기반 FTL 캐시 분리 기법)

  • Bae, Jinwook;Kim, Hanbyeol;Im, Junsu;Lee, Sungjin
    • IEMEK Journal of Embedded Systems and Applications
    • /
    • v.14 no.2
    • /
    • pp.71-78
    • /
    • 2019
  • As the capacity of SSDs rapidly increases, the amount of DRAM to keep a mapping table size in SSDs becomes very huge. To address a Demand-based FTL (DFTL) scheme that caches part of mapping entries in DRAM is considered to be a feasible alternative. However, owing to its unpredictable behaviors, DFTL fails to provide consistent I/O response times. In this paper, we a) analyze a root cause that results in fluctuation on read latency and b) propose a new demand-based FTL scheme that ensures guaranteed read response time with low write amplification. By preventing mapping evictions while serving reads, the proposed technique guarantees every host read requests to be done in 2 NAND read operations. Moreover, only with 25% of a cache ratio, the proposed scheme improves random write performance and random mixed performance by 1.65x and 1.15x, respectively, over the traditional DFTL.

WWCLOCK: Page Replacement Algorithm Considering Asymmetric I/O Cost of Flash Memory (WWCLOCK: 플래시 메모리의 비대칭적 입출력 비용을 고려한 페이지 교체 알고리즘)

  • Park, Jun-Seok;Lee, Eun-Ji;Seo, Hyun-Min;Koh, Kern
    • Journal of KIISE:Computing Practices and Letters
    • /
    • v.15 no.12
    • /
    • pp.913-917
    • /
    • 2009
  • Flash memories have asymmetric I/O costs for read and write in terms of latency and energy consumption. However, the ratio of these costs is dependent on the type of storage. Moreover, it is becoming more common to use two flash memories on a system as an internal memory and an external memory card. For this reason, buffer cache replacement algorithms should consider I/O costs of device as well as possibility of reference. This paper presents WWCLOCK(Write-Weighted CLOCK) algorithm which directly uses I/O costs of devices along with recency and frequency of cache blocks to selecting a victim to evict from the buffer cache. WWCLOCK can be used for wide range of storage devices with different I/O cost and for systems that are using two or more memory devices at the same time. In addition to this, it has low time and space complexity comparable to CLOCK algorithm. Trace-driven simulations show that the proposed algorithm reduces the total I/O time compared with LRU by 36.2% on average.

FlashEDF: An EDF-style Scheduling Scheme for Serving Real-time I/O Requests in Flash Storage

  • Lim, Seong-Chae
    • International Journal of Internet, Broadcasting and Communication
    • /
    • v.10 no.3
    • /
    • pp.26-34
    • /
    • 2018
  • In this paper, we propose a scheduling scheme that can efficiently serve I/O requests having deadlines in flash storage. The I/O requests with deadlines, namely, real-time requests, are assumed to be issued for streaming services of continuous media. Since a Web-based streaming server commonly supports downloads of HTMLs or images, we also aim to quickly process non-real-time I/O requests, together with real-time ones. For this purpose, we adopt the well-known rate-reservation EDF (RR-EDF) algorithm for determining scheduling priorities among mixed I/O requests. In fact, for the use of an EDF-style algorithm, overhead of task's switching should be low and predictable, as with its application of CPU scheduling. In other words, the EDF algorithm is inherently unsuitable for scheduling I/O requests in HDD storage because of highly varying latency times of HDD. Unlike HDD, time for reading a block in flash storage is almost uniform with respect to its physical location. This is because flash storage has no mechanical component, differently from HDD. By capitalizing on this uniform block read time, we compute bandwidth utilization rates of real-time requests from streams. Then, the RR-EDF algorithm is applied for determining how much storage bandwidth can be assigned to non-real-time requests, while meeting deadlines of real-time requests. From this, we can improve the service times of non-real-time requests, which are issued for downloads of static files. Because the proposed scheme can expand flexibly the scheduling periods of streams, it can provide a full usage of slack times, thereby improving the overall throughput of flash storage significantly.

Design and Evaluation of an Adaptive Reservation-based Batching Polity for Popular Videos (인기 있는 비디오를 위한 적응적 예약기반 일괄처리 정책의 설계 및 평가)

  • Lee, Gyeong-Suk;Bae, In-Han
    • The Transactions of the Korea Information Processing Society
    • /
    • v.6 no.10
    • /
    • pp.2790-2796
    • /
    • 1999
  • In video-on-demand systems, the I.O bandwidth of video servers is the critical resource which contributes to increase in latency. Several approaches: bridging, piggybacking are used to reduce the I/O demand on the video server through sharing. Batching delays the requests for the different videos for a batching window so that more requests for the same video arriving during the current batching window may be served using the same stream. In this paper, we propose an adaptive reservation-based batching policy which dynamically reserves video server capacity for popoular videos according to video server loads. The performance of the proposed policy is evaluated through a simulation, and is compared with simple batching and static reservation-based batching policies. As the result, we know that the adaptive reservation-based batching policy more improves service ratio and average waiting time than simple batching and simple reservation-based batching policy more improves service ratio and average waiting time than simple batching and simple reservation-based batching polices.

  • PDF

Pig Skin Gelatin Hydrolysates Attenuate Acetylcholine Esterase Activity and Scopolamine-induced Impairment of Memory and Learning Ability of Mice

  • Kim, Dongwook;Kim, Yuan H. Brad;Ham, Jun-Sang;Lee, Sung Ki;Jang, Aera
    • Food Science of Animal Resources
    • /
    • v.40 no.2
    • /
    • pp.183-196
    • /
    • 2020
  • The protective effect of pig skin gelatin water extracts (PSW) and the low molecular weight hydrolysates of PSW generated via enzymatic hydrolysis with Flavourzyme® 1000L (LPSW) against scopolamine-induced impairment of cognitive function in mice was determined. Seventy male ICR mice weighing 20-25 g were randomly assigned to seven groups: Control (CON); scopolamine (SCO, 1 mg/kg B.W., intraperitoneally (i.p.); tetrahydroaminoacridine 10 [THA 10, tacrine; 10 mg/kg B.W. per oral (p.o.) with SCO (i.p.)]; PSW 10 (10 mg/kg B.W. (p.o.) with SCO (i.p.); PSW 40 (40 mg/kg B.W. (p.o.) with SCO (i.p.); LPSW 100 (100 mg/kg B.W. (p.o.) with SCO (i.p.); LPSW 400 (400 mg/kg B.W. (p.o.) with SCO (i.p.). All treatment groups, except CON, received scopolamine on the day of the experiment. The oxygen radical absorbance capacity of LPSW 400 at 1 mg/mL was 154.14 μM Trolox equivalent. Administration of PSW and LPSW for 15 weeks did not significantly affect on physical performance of mice. LPSW 400 significantly increased spontaneous alternation, reaching the level observed for THA and CON. The latency time of animals receiving LPSW 400 was higher than that of mice treated with SCO alone in the passive avoidance test, whereas it was shorter in the water maze test. LPSW 400 increased acetylcholine (ACh) content and decreased ACh esterase activity (p<0.05). LPSW 100 and LPSW 400 reduced monoamine oxidase-B activity. These results indicated that LPSW at 400 mg/kg B.W. is a potentially strong antioxidant and contains novel components for the functional food industry.

Performance Analysis of Virtual Storage

  • Wei, X;RO, CheulWoo
    • Proceedings of the Korea Contents Association Conference
    • /
    • 2018.05a
    • /
    • pp.511-512
    • /
    • 2018
  • Virtual machine technology enables multiple hosts to share the same LUN(Logical Unit Number) and the same storage, but if too many hosts share the same LUN, it will increase the delay. In this paper we propose a performance model, get several values of scalable storage performance in virtual environment, and present the results examined the effects of I/O queuing in a virtual infrastructure. This results show how to make the most effective use of our storage resources.

  • PDF

QEMU/KVM Based In-Memory Block Cache Module for Virtualization Environment (가상화 환경을 위한 QEMU/KVM 기반의 인메모리 블록 캐시 모듈 구현)

  • Kim, TaeHoon;Song, KwangHyeok;No, JaeChun;Park, SungSoon
    • Journal of KIISE
    • /
    • v.44 no.10
    • /
    • pp.1005-1018
    • /
    • 2017
  • Recently, virtualization has become an essential component of cloud computing due to its various strengths, including maximizing server resource utilization, easy-to-maintain software, and enhanced data protection. However, since virtualization allows sharing physical resources among the VMs, the system performance can be deteriorated due to device contentions. In this paper, we first investigate the I/O overhead based on the number of VMs on the same server platform and analyze the block I/O process of the KVM hypervisor. We also propose an in-memory block cache mechanism, called QBic, to overcome I/O virtualization latency. QBic is capable of monitoring the block I/O process of the hypervisor and stores the data with a high access frequency in the cache. As a result, QBic provides a fast response for VMs and reduces the I/O contention to physical devices. Finally, we present a performance measurement of QBic to verify its effectiveness.