• Title/Summary/Keyword: Cache Hit Ratio

Search Result 107, Processing Time 0.028 seconds

Block Replacement Scheme based on Reuse Interval for Hybrid SSD System (Hybrid SSD 시스템을 위한 재사용 간격 기반 블록 교체 기법)

  • Yoo, Sanghyun;Kim, Kyung Tae;Youn, Hee Yong
    • Journal of Internet Computing and Services
    • /
    • v.16 no.5
    • /
    • pp.19-27
    • /
    • 2015
  • Due to the advantages of fast read/write operation and low power consumption, SSD(Solid State Drive) is now widely adopted as storage device of smart phone, laptop computer, server, etc. However, the shortcomings of SSD such as limited number of write operations and asymmetric read/write operation lead to the problem of shortened life span of SSD. Therefore, the block replacement policy of SSD used as cache for HDD is very important. The existing solutions for improving the lifespan of SSD including the LARC scheme typically employ the LRU algorithm to manage the SSD blocks, which may increase the miss rate in SSD due to the replacement of frequently used block instead of rarely used block. In this paper we propose a novel block replacement scheme which considers the block reuse interval to effectively handle various data read/write patterns. The proposed scheme replaces the block in SSD based on the recency decided by reuse interval and age along with hit ratio. Computer simulation using workload trace files reveals that the proposed scheme consistently improves the performance and lifespan of SSD by increasing the hit ratio and decreasing the number of write operations compared to the existing schemes including LARC.

A novel page replacement policy associated with ACT-R inspired by human memory retrieval process (인간 기억 인출 과정을 응용하여 설계된 ACT-R 기반 페이지 교체 정책)

  • Roh, Hong-Chan;Park, Sang-Hyun
    • The KIPS Transactions:PartD
    • /
    • v.18D no.1
    • /
    • pp.1-8
    • /
    • 2011
  • The cache structure, which is designed for assuring fast accesses to frequently accessed data, resides on the various levels of computer system hierarchies. Many studies on this cache structure have been conducted and thus many page-replacement algorithms have been proposed. Most of page-replacement algorithms are designed on the basis of heuristic methods by using their own criteria such as how recently pages are accessed and how often they are accessed. This data-retrieval process in computer systems is analogous to human memory retrieval process since the retrieval process of human memory depends on frequency and recency of the retrieval events as well. A recent study regarding human memory cognition revealed that the possibility of the retrieval success and the retrieval latency have a strong correlation with the frequency and recency of the previous retrieval events. In this paper, we propose a novel page-replacement algorithm by utilizing the knowledge from the recent research regarding human memory cognition. Through a set of experiments, we demonstrated that our new method presents better hit-ratio than the LRFU algorithm which has been known as the best performing page-replacement algorithm for DBMS caches.

Performance Evaluation of Caching in PON-based 5G Fronthaul (PON기반 5G 프론트홀의 캐싱 성능 평가)

  • Jung, Bokrae
    • Journal of Convergence for Information Technology
    • /
    • v.10 no.1
    • /
    • pp.22-27
    • /
    • 2020
  • With the deployment of 5G infrastructure, content delivery network (CDN) will be a key role to provide explosive growing services for the independent media and YouTube which contain high-speed mobile contents. Without a local cache, the mobile backhaul and fronthaul should endure huge burden of bandwidth request for users as the increase number of direct accesses from contents providers. To deal with this issue, this paper fist presents both fronthaul solutions for CDN that use dark fibers and a passive optical network (PON). On top of that, we propose the aggregated content request specialized for PON caching and evaluate and compare its performance to legacy schemes through the simulation. The proposed PON caching scheme can reduce average access time of up to 0.5 seconds, 1/n received request packets, and save 60% of backhaul bandwidth compared to the no caching scheme. This work can be a useful reference for service providers and will be extended to further improve the hit ratio of cache in the future.

Interchange Algorithm for VoD System (VOD 시스템에서의 Interchange Agent 운영 알고리즘)

  • Kang, Seok-Hoon;Park, Su-Hyun
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.9 no.8
    • /
    • pp.1847-1854
    • /
    • 2005
  • This paper proposes a approach to configure efficient video-on-demand system by introducing Multicast and Cache Video-on-Demand (MCVoD) system. As a key element or the MCVoD system, interchange agent provides this system with multicasting and switching functions. With the multicasting, the MCVoD system is able to reduce the load on the network as well as VoD servers by transmitting only one video request instead of sending multiple requests on a same video stream. The switching enables clients to receive the lust stream of requested video streams instantly without waiting time and also allows avoiding undesirable duplication of video streams in the system. With various experiment results through simulation about waiting tine and cache hit ratio, we show that the MCVoD system employing the interchange agent provides better performance than current uni-proxy based system.

Improving Hit Ratio and Hybrid Branch Prediction Performance with Victim BTB (Victim BTB를 활용한 히트율 개선과 효율적인 통합 분기 예측)

  • Joo, Young-Sang;Cho, Kyung-San
    • The Transactions of the Korea Information Processing Society
    • /
    • v.5 no.10
    • /
    • pp.2676-2685
    • /
    • 1998
  • In order to improve the branch prediction accuracy and to reduce the BTB miss rate, this paper proposes a two-level BTB structure that adds small-sized victim BTB to the convetional BTB. With small cost, two-level BTB can reduce the BTB miss rate as well as improve the prediction accuracy of the hybrid branch prediction strategy which combines dynamic prediction and static prediction. Through the trace-driven simulation of four bechmark programs, the performance improvement by the proposed two-level BTB structure is analysed and validated. Our proposed BTB structure can improve the BTB miss rate by 26.5% and the misprediction rate by 26.75%

  • PDF

Social-Aware Collaborative Caching Based on User Preferences for D2D Content Sharing

  • Zhang, Can;Wu, Dan;Ao, Liang;Wang, Meng;Cai, Yueming
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.14 no.3
    • /
    • pp.1065-1085
    • /
    • 2020
  • With rapid growth of content demands, device-to-device (D2D) content sharing is exploited to effectively improve the service quality of users. Considering the limited storage space and various content demands of users, caching schemes are significant. However, most of them ignore the influence of the asynchronous content reuse and the selfishness of users. In this work, the user preferences are defined by exploiting the user-oriented content popularity and the current caching situation, and further, we propose the social-aware rate, which comprehensively reflects the achievable contents download rate affected by the social ties, the caching indicators, and the user preferences. Guided by this, we model the collaborative caching problem by making a trade-off between the redundancy of caching contents and the cache hit ratio, with the goal of maximizing the sum of social-aware rate over the constraint of limited storage space. Due to its intractability, it is computationally reduced to the maximization of a monotone submodular function, subject to a matroid constraint. Subsequently, two social-aware collaborative caching algorithms are designed by leveraging the standard and continuous greedy algorithms respectively, which are proved to achieve different approximation ratios in unequal polynomial-time. We present the simulation results to illustrate the performance of our schemes.

Object Segment Grouping for Wireless Mobile Streaming Media Services (무선 모바일 스트리밍 미디어 서비스를 위한 객체 세그먼트 그룹화)

  • Lee, Chong-Deuk
    • Journal of Digital Convergence
    • /
    • v.10 no.4
    • /
    • pp.199-206
    • /
    • 2012
  • Increment of mobile client's information request in wireless mobile networks requires a new method to manage and serve the streaming media object. This paper proposes a new object segment grouping method for enhancing the performance of streaming media services in wireless mobile networks. The proposed method performs the similarity metric for the partitioned object segments, and it process the disjunction, conjunction, and filtering for these metrics. This paper was to decided the partitioned group of object segments for these operation metrics, and it decided the performance of streaming media services. The simulation result showed that the proposed method has better performance in throughput, average startup latency, and cache hit ratio.

Extended Buffer Management with Flash Memory SSDs (플래시메모리 SSD를 이용한 확장형 버퍼 관리)

  • Sim, Do-Yoon;Park, Jang-Woo;Kim, Sung-Tan;Lee, Sang-Won;Moon, Bong-Ki
    • Journal of KIISE:Databases
    • /
    • v.37 no.6
    • /
    • pp.308-314
    • /
    • 2010
  • As the price of flash memory continues to drop and the technology of flash SSD controller innovates, high performance flash SSDs with affordable prices flourish in the storage market. Nevertheless, it is hard to expect that flash SSDs will replace harddisks completely as database storage. Instead, the approach to use flash SSD as a cache for harddisks would be more practical, and, in fact, several hybrid storage architectures for flash memory and harddisk have been suggested in the literature. In this paper, we propose a new approach to use flash SSD as an extended buffer for main buffer in database systems, which stores the pages replaced out from main buffer and returns the pages which are re-referenced in the upper buffer layer, improving the system performance drastically. In contrast to the existing approaches to use flash SSD as a cache in the lower storage layer, our approach, which uses flash SSD as an extended buffer in the upper host, can provide fast random read speed for the warm pages which are being replaced out from the limited main buffer. In fact, for all the pages which are missing from the main buffer in a real TPC-C trace, the hit ratio in the extended buffer could be more than 60%, and this supports our conjecture that our simple extended buffer approach could be very effective as a cache. In terms of performance/price, our extended buffer architecture outperforms two other alternative approaches with the same cost, 1) large main buffer and 2) more harddisks.

A Performance Improvement of Linux TCP/IP Stack based on Flow-Level Parallelism in a Multi-Core System (멀티코어 시스템에서 흐름 수준 병렬처리에 기반한 리눅스 TCP/IP 스택의 성능 개선)

  • Kwon, Hui-Ung;Jung, Hyung-Jin;Kwak, Hu-Keun;Kim, Young-Jong;Chung, Kyu-Sik
    • The KIPS Transactions:PartA
    • /
    • v.16A no.2
    • /
    • pp.113-124
    • /
    • 2009
  • With increasing multicore system, much effort has been put on the performance improvement of its application. Because multicore system has multiple processing devices in one system, its processing power increases compared to the single core system. However in many cases the advantages of multicore can not be exploited fully because the existing software and hardware were designed to be suitable for single core. When the existing software runs on multicore, its performance improvement is limited by the bottleneck of sharing resources and the inefficient use of cache memory on multicore. Therefore, according as the number of core increases, it doesn't show performance improvement and shows performance drop in the worst case. In this paper we propose a method of performance improvement of multicore system by applying Flow-Level Parallelism to the existing TCP/IP network application and operating system. The proposed method sets up the execution environment so that each core unit operates independently as much as possible in network application, TCP/IP stack on operating system, device driver, and network interface. Moreover it distributes network traffics to each core unit through L2 switch. The proposed method allows to minimize the sharing of application data, data structure, socket, device driver, and network interface between each core. Also it allows to minimize the competition among cores to take resources and increase the hit ratio of cache. We implemented the proposed methods with 8 core system and performed experiment. Experimental results show that network access speed and bandwidth increase linearly according to the number of core.

An Efficient Cooperative Web Caching Scheme (효율적인 협동적 웹캐슁 기법)

  • Shin, Yong-Hyeon
    • The KIPS Transactions:PartC
    • /
    • v.13C no.6 s.109
    • /
    • pp.785-794
    • /
    • 2006
  • Nowadays, Internet is used worldwide and network traffic is increasing dramatically. Much of Internet traffic is due to the web applications. And I propose a new cooperative web caching scheme, called DCOORD which tries to minimize the overall cost of Web caching. DCOORD reduces the communication cost by coordinating the objects which are cached at each cache server. In this paper, I compare the Performance of DCOORD with two well-known cooperative Web caching schemes, ICP and CARP, using trace driven simulation. In order to reflect the cost factor in the network communication, I used the CSR(Cost-Saving Ratio) as our performance metric, instead of the traditional hit ratio. The performance evaluations show that DCOORD is more cost effective than ICP and CARP.