• Title/Summary/Keyword: Internal Caching

Search Result 8, Processing Time 0.021 seconds

An performance analysis on SSD caching mechanism in Linux (리눅스 SSD caching mechanism 의 성능 비교 및 분석)

  • Heo, Sang-Bok;Park, Jinhee;Jo, Heeseung
    • Smart Media Journal
    • /
    • v.4 no.2
    • /
    • pp.62-67
    • /
    • 2015
  • During several decades, hard disk drive(HDD) has been used in most computer systems as secondary storage and, however, the performance enhancement of HDD is limited by its mechanical properties. On the other hand, although the flash memory based solid state drive (SSD) has more advantages over HDD such as high performance and low noise, SSD is still too expensive for common usage and expected to take several years to replace HDD completely. Therefore, SSD caching mechanism using the SSD as a cache of high capacity HDD has been highlighted lately. The representatives of SSD caching mechanisms are typically bcache, dm-cache, Flashcache, and EnhanceIO. Each of them has its own internal mechanism and implementation, and this makes them to show their own pros. and cons. In this paper, we analyze the characteristics of each SSD caching mechanisms and compare the performance of them under various workloads. We expect that our contribution will be useful to enhance the performance of SSD caching mechanisms.

Fast Multi-Phase Packet Classification Architecture using Internal Buffer and Single Entry Caching (내부 버퍼와 단일 엔트리 캐슁을 이용한 다단계 패킷 분류 가속화 구조)

  • Kang, Dae-In;Park, Hyun-Tae;Kim, Hyun-Sik;Kang, Sung-Ho
    • Journal of the Institute of Electronics Engineers of Korea TC
    • /
    • v.44 no.9
    • /
    • pp.38-45
    • /
    • 2007
  • With the emergence of new applications, packet classification is essential for supporting advanced internet applications, such as network security and QoS provisioning. As the packet classification on multiple-fields is a difficult and time consuming problem, internet routers need to classify incoming packet quickly into flows. In this paper, we present multi-phase packet classification architecture using an internal buffer for fast packet processing. Using internal buffer between address pair searching phase and remained fields searching phases, we can hide latency from the characteristic that search times of source and destination header fields are different. Moreover we guarantee the improvement by using single entry caching. The proposed architecture is easy to apply to different needs owing to its simplicity and generality.

A P2P Based Tactical Information Sharing System for Mobile Nodes (P2P 기반의 모바일 노드간의 전술 정보 공유 시스템)

  • Lee, Hyun
    • Journal of the Korea Institute of Military Science and Technology
    • /
    • v.17 no.4
    • /
    • pp.501-509
    • /
    • 2014
  • In NCW(Network Centric Warfare) environment, mobile nodes communicate through wireless link. But wireless link provides limited networking performance due to signal interferences or mobility of nodes. So it is quite challenge to acquire enough networking resources and use the resources efficiently. In this paper, we have proposed a P2P based tactical information sharing system which provides satisfactory visual information playout for mobile nodes(i.e., military personnel, vehicle,..) in NCW environment. Our proposed system consists of two components. One is caching-enabled switch which stores tactical information segments at its internal storage and then transports them to mobile nodes when require. Another is centralized scheduling algorithm which exploits networking resources more efficiently. To validate performance of proposed system, we performed series of experiments in wireless network testbed. Results show improved performance in terms of segment-missing ratio, networking resources usage, sharing time, and number of simultaneous playout mobile nodes with acceptable playout continuity(i.e., over 95%).

Parallel Multithreaded Processing for Data Set Summarization on Multicore CPUs

  • Ordonez, Carlos;Navas, Mario;Garcia-Alvarado, Carlos
    • Journal of Computing Science and Engineering
    • /
    • v.5 no.2
    • /
    • pp.111-120
    • /
    • 2011
  • Data mining algorithms should exploit new hardware technologies to accelerate computations. Such goal is difficult to achieve in database management system (DBMS) due to its complex internal subsystems and because data mining numeric computations of large data sets are difficult to optimize. This paper explores taking advantage of existing multithreaded capabilities of multicore CPUs as well as caching in RAM memory to efficiently compute summaries of a large data set, a fundamental data mining problem. We introduce parallel algorithms working on multiple threads, which overcome the row aggregation processing bottleneck of accessing secondary storage, while maintaining linear time complexity with respect to data set size. Our proposal is based on a combination of table scans and parallel multithreaded processing among multiple cores in the CPU. We introduce several database-style and hardware-level optimizations: caching row blocks of the input table, managing available RAM memory, interleaving I/O and CPU processing, as well as tuning the number of working threads. We experimentally benchmark our algorithms with large data sets on a DBMS running on a computer with a multicore CPU. We show that our algorithms outperform existing DBMS mechanisms in computing aggregations of multidimensional data summaries, especially as dimensionality grows. Furthermore, we show that local memory allocation (RAM block size) does not have a significant impact when the thread management algorithm distributes the workload among a fixed number of threads. Our proposal is unique in the sense that we do not modify or require access to the DBMS source code, but instead, we extend the DBMS with analytic functionality by developing User-Defined Functions.

An Efficient Data Nigration/Replication Scheme in a Large Scale Multimedia Server (대규모 멀티미디어 서버에서 효율적인 데이터 이동/중복 기법)

  • Kim, Eun-Sam
    • Journal of the Korea Society of Computer and Information
    • /
    • v.14 no.5
    • /
    • pp.37-44
    • /
    • 2009
  • Recently, as the quality of multimedia data gets higher, multimedia servers require larger storage capacity and higher I/O bandwidth. In these large scale multimedia servers, the load-unbalance problem among disks due to the difference in access frequencies to multimedia objects according to their popularities significantly affects the system performance. To address this problem, many data replication schemes have been proposed. In this paper, we propose a novel data migration/replication scheme to provide better storage efficiency and performance than the dynamic data replication scheme which is typical data replication scheme employed in multimedia servers. This scheme can reduce the additional storage space required for replication, which is a major defect of replication schemes, by decreasing the number of copies per object. The scheme can also increase the number of concurrent users by increasing the caching effect due to the reduced lengths of the intervals among requests for each object.

Implementation of Memory Efficient Flash Translation Layer for Open-channel SSDs

  • Oh, Gijun;Ahn, Sungyong
    • International journal of advanced smart convergence
    • /
    • v.10 no.1
    • /
    • pp.142-150
    • /
    • 2021
  • Open-channel SSD is a new type of Solid-State Disk (SSD) that improves the garbage collection overhead and write amplification due to physical constraints of NAND flash memory by exposing the internal structure of the SSD to the host. However, the host-level Flash Translation Layer (FTL) provided for open-channel SSDs in the current Linux kernel consumes host memory excessively because it use page-level mapping table to translate logical address to physical address. Therefore, in this paper, we implemente a selective mapping table loading scheme that loads only a currently required part of the mapping table to the mapping table cache from SSD instead of entire mapping table. In addition, to increase the hit ratio of the mapping table cache, filesystem information and mapping table access history are utilized for cache replacement policy. The proposed scheme is implemented in the host-level FTL of the Linux kernel and evaluated using open-channel SSD emulator. According to the evaluation results, we can achieve 80% of I/O performance using the only 32% of memory usage compared to the previous host-level FTL.

I/O Scheme of Hybrid Hard Disk Drive for Low Power Consumption and Effective Response Time (저전력과 응답시간 향상을 위한 하이브리드 하드디스크의 입출력 기법)

  • Kim, Jeong-Won
    • Journal of the Korea Society of Computer and Information
    • /
    • v.16 no.10
    • /
    • pp.23-31
    • /
    • 2011
  • Recently, Solid state disk is mainly used because this device has lower power consumption as well as higher response time. But it features higher price and lower performance at delete and write operations compared with HDD. To compensate this defect, Hybrid hard disk with internal non-volatile flash memory was issued. This NVCache is used as a kind of cache for disk blocks. In this paper, an I/O scheme for H-HDD is proposed for improving low power consumption as well as response time. Our method is to use this NVCache as read cache mainly and write cache when write requests are concentrated. In read cache operation, disk blocks with higher priority determined on basis of time as well as spatial localities are prefetched, which can improve response time. The write operation is conducted only at write peak time as disk spindle up costs higher battery power as well as response time. Experiments results show that the suggested method can improve response time of H-HDD and lower the power consumption.

Analysis and Advice on Cache Algorithms of SSD FTL (SSD FTL 캐시 알고리즘 분석 및 제언)

  • Hyung Bong, Lee;Tae Yun, Chung
    • KIPS Transactions on Computer and Communication Systems
    • /
    • v.12 no.1
    • /
    • pp.1-8
    • /
    • 2023
  • It is impossible to overwrite on an already allocated page in SSDs, so whenever a write operation occurs a page replacement with a clean page is required. To resolve this problem, SSDs have an internal flash translation layer called FTL that maps logical pages managed by a file system of operating system to currently allocated physical pages. SSD pages discarded due to write operations must be recycled through initialization, but since the number of initialization times is limited the FTL provides a caching function to reduce the number of writes in addition to the page mapping function, which is a core function. In this study, we focus on the FTL cache methodologies reducing the number of page writes and analyze the related algorithms, and propose a write-only cache strategy. As a result of experimenting with the write-only cache using a simulator, it showed an improvement of up to 29%.