• Title/Summary/Keyword: LFU

Search Result 41, Processing Time 0.024 seconds

MLC-LFU : The Multi-Level Buffer Cache Management Policy for Flash Memory (MLC-LFU : 플래시 메모리를 위한 멀티레벨 버퍼 캐시 관리 정책)

  • Ok, Dong-Seok;Lee, Tae-Hoon;Chung, Ki-Dong
    • Journal of KIISE:Computing Practices and Letters
    • /
    • v.15 no.1
    • /
    • pp.14-20
    • /
    • 2009
  • Recently, NAND flash memory is used not only for portable devices, but also for personal computers and server computers. Buffer cache replacement policies for the hard disks such as LRU and LFU are not good for NAND flash memories because they do not consider about the characteristics of NAND flash memory. CFLRU and its variants, CFLRU/C, CFLRU/E and DL-CFLRU/E(CFLRUs) are the buffer cache replacement policies considered about the characteristics of NAND flash memories, but their performances are not better than those of LRD. In this paper, we propose a new buffer cache replacement policy for NAND flash memory. Which is based on LFU and is taking into account the characteristics of NAND flash memory. And we estimate the performance of hit ratio and flush operation numbers. The proposed policy shows better hit ratio and the number of flush operation than any other policies.

Delay Attenuation LFU (DA-LFU) Cache Replacement Policy to Improve Hit Rates in CCN (CCN에서 적중률 향상을 위한 지연감쇠 LFU(DA-LFU) 캐시 교체 정책)

  • Ban, Bin;Kwon, Tae-Wook
    • KIPS Transactions on Computer and Communication Systems
    • /
    • v.9 no.3
    • /
    • pp.59-66
    • /
    • 2020
  • Content Centric Network(CCN) with architecture that is completely different from traditional host-based networks has emerged to address problems such as the explosion of traffic load in the current network. Research on cache replacement policies is very active to improve the performance of CCN with the characteristics that all routers cache on the network. Therefore, this paper proposes a cache replacement policy suitable for situations in which popularity is constantly changing, taking into account the actual network situation. In order to evaluate the performance of the proposed algorithm, we experimented in an environment where the popularity of content is constantly changing, and confirmed that we are superior to the existing replacement policy through comparing hit rates and analyzing server load.

A LFU based on Real-time Producer Popularity in Concent Centric Networks (CCN에서 실시간 생성자 인기도 기반의 LFU 정책)

  • Choi, Jong-Hyun;Kwon, Tea-Wook
    • The Journal of the Korea institute of electronic communication sciences
    • /
    • v.16 no.6
    • /
    • pp.1113-1120
    • /
    • 2021
  • Content Central Network (CCN) appeared to improve network efficiency by transforming IP-based network into content name-based network structures. Each router performs caching mechanism to improve network efficiency in the CCN. And the cache replacement policy applied to the CCN router is an important factor that determines the overall performance of the CCN. Therefore various studies has been done relating to cache replacement policy of the CCN. In this paper, we proposed a cache replacement policy that improves the limitations of the LFU policy. The proposal algorithm applies real-time producer popularity-based variables. And through experiments, we proved that the proposed policy shows a better cache hit ratio than existing policies.

On the study of block replacement policy using LFR (LFR기법을 이용한 블럭교체 기법에 관한 연구)

  • 오재환;김상수김미선
    • Proceedings of the IEEK Conference
    • /
    • 1998.10a
    • /
    • pp.499-502
    • /
    • 1998
  • Most popular disk block replacement polices are LRU(Least Recently Used)policy and LFU(Least Frequently Used)policy. The LRU policy replaces blocks according to the most recent reference without considering the frequency of reference. The LFU policy replaces blocks according to the frequency of reference without considering the recently of the reference. In this thesis, a policy called LFR(least Frequently Use & Not Used Recently) disk block replacement policy is presented. The LFR policy subsumes the LFU policy and the NUR policy. The spectrum of the LFR policy exists between the LFU policy and the NUR policy because we co give different weight to each reference of a block. The implementation shows LFR policy outperforms the previously implemente LRU policy.

  • PDF

Memory Reference Patterns and Page Replacement Policies (메모리 참조 패턴 및 페이지 교체 기법)

  • 이승훈;최종무;조성제;조유근
    • Proceedings of the Korean Information Science Society Conference
    • /
    • 2000.10c
    • /
    • pp.50-52
    • /
    • 2000
  • 페이징 기법의 효율성은 어떠한 페이지 교체 기법을 쓰느냐에 따라 결정된다. LRU 기법은 작업 부하의 변화에 잘 적응하여 많은 경우 좋은 성능을 보인다. 그러나 참조의 횟수에 대한 정보를 이용하지 못한다. LFU 기법은 지역성을 가지는 참조 패턴이 발견되면 좋은 성능을 발휘한다. 그러나 작업 부하가 변하는 경우 이에 적응하지 못한다. 여러 응용에 대해 참조 패턴을 분석하여 보면 참조되는 페이지는 최근성과 참조 횟수에 의해 가치가 결정되며, 따라서 LRU나 LFU 기법 한 가지만으로 페이지 교체 정책을 최적화 시킬 수 없다. 본 논문에서는 LRU 기법과 LFU 기법을 결합한 새로운 교체 기법을 제안한다. 이 교체 기법에서는 LRU 리스트와 LFU 리스트를 결합하여 사용함으로써 참조 시간뿐만 아니라 참조 횟수를 이용하여 페이지들을 교체한다. 트레이스 기반 모의 실험에서는 제안 기법의 순수 LRU 기법보다 나은 성능을 보일 때가 있다.

  • PDF

A Page Replacement Scheme Based on Recency and Frequency (최근성과 참조 횟수에 기반한 페이지 교체 기법)

  • Lee, Seung-Hoon;Lee, Jong-Woo;Cho, Seong-Je
    • The KIPS Transactions:PartA
    • /
    • v.8A no.4
    • /
    • pp.469-478
    • /
    • 2001
  • In the virtual memory system, page replacement policy exerts a great influence on the performance of demand paging. There are LRU(Least Recently Used) and LFU (Least Frequently Used) as the typical replacement policies. The LRU policy performs effectively in many cases and adapts well to the changing workloads compared to other policies. It however cannot distinguish well between frequently and infrequently referenced pages. The LFU policy requires that the page with the smallest reference count be replaced. Though it considers all the references in the past, it cannot discriminate between references that occurred far back in the past and the more recent ones. Thus, it cannot adapt well to the changing workload. In this paper, we first analyze memory reference patterns of eight applications. The patterns show that the recently referenced pages or the frequently referenced pages are accessed continuously as the case may be. So it is rather hard to optimize page replacement scheme by using just one of the LRU or LFU policy. This paper makes an attempt to combine the advantages of the two policies and proposes a new page replacement policy. In the proposed policy, paging list is divided into two lists (LRU and LFU lists). By keeping the two lists in recency and reference frequency order respectively, we try to restrain the highly referenced pages in the past from being replaced by the LRU policy. Results from trace-driven simulations show that there exists points on the spectrum at which the proposed policy performs better than the previously known policies for the workloads we considered. Especially, we can see that our policy outperforms the existing ones in such applications that have reference patterns of re-accessing the frequently referenced pages in the past after some time.

  • PDF

A Caching Mechanism for Remote Queries in Distributed Directory Systems (분산 디렉토리 시스템에서의 원격 질의에 대한 캐싱 기법)

  • Lee, Kang-Woo
    • The Transactions of the Korea Information Processing Society
    • /
    • v.7 no.1
    • /
    • pp.50-56
    • /
    • 2000
  • In this paper, for improving the speed of query processing on distributed directory system, we proposed a caching mechanism which is store the queries and their on the remote site objects in the cache of local site. For this, first, cached information which is stored in distributed directory systems is classified as application data and system data. And cache system architecture is designed according to classified information. Second least-TTL replacement mechanism which uses the weighted value of geographical information and access frequency for replacements are developed for each cache. Finally, performance evaluations are performed by comparing the proposed caching mechanism and other mechanisms (LRU, LFU replacement). Our least-TTL mechanism shows a performance improvement of 25% over the LRU and that of 30% over LFU.

  • PDF

An Efficient Data Block Replacement and Rearrangement Technique for Hybrid Hard Disk Drive (하이브리드 하드디스크를 위한 효율적인 데이터 블록 교체 및 재배치 기법)

  • Park, Kwang-Hee;Lee, Geun-Hyung;Kim, Deok-Hwan
    • Journal of KIISE:Computing Practices and Letters
    • /
    • v.16 no.1
    • /
    • pp.1-10
    • /
    • 2010
  • Recently heterogeneous storage system such as hybrid hard disk drive (H-HDD) combining flash memory and magnetic disk is launched, according as the read performance of NAND flash memory is enhanced as similar to that of hard disk drive (HDD) and the power consumption of NAND flash memory is reduced less than that of HDD. However, the read and write operations of NAND flash memory are slower than those of rotational disk. Besides, serious overheads are incurred on CPU and main memory in the case that intensive write requests to flash memory are repeatedly occurred. In this paper, we propose the Least Frequently Used-Hot scheme that replaces the data blocks whose reference frequency of read operation is low and update frequency of write operation is high, and the data flushing scheme that rearranges the data blocks into the multi-zone of the rotation disk. Experimental results show that the execution time of the proposed method is 38% faster than those of conventional LRU and LFU block replacement schemes in I/O performance aspect and the proposed method increases the life span of Non-Volatile Cache 40% higher than those of conventional LRU, LFU, FIFO block replacement schemes.

Cache Replacement Policies for Energy Efficiency (저전력 캐쉬 대체 정책)

  • 이문상;이원진;이준원;맹승렬
    • Proceedings of the Korean Information Science Society Conference
    • /
    • 2000.04a
    • /
    • pp.12-14
    • /
    • 2000
  • 소형의 이동 컴퓨터 시스템이 발전하면서 프로세서의 전력 소모(energy dissipation)가 중요한 이슈가 되고 있다. 현재 대부분의 프로세서들은 성능 향상을 위해 캐쉬를 사용하고 있고 이것은 프로세서내의 많은 비율의 전력을 소모한다. 따라서 저 전력 프로세서를 설계하기 위해서는 내장 캐쉬(on-chip cache)의 전력 소모를 줄이는 것이 중요하다. 본 논문은 캐쉬 대체 전략으로 현재 많이 사용되는 LRU(Least Recently Used) 방식을 LFU(Least Frequently Used), LFUT(LFU with Threshold), FIFO(First In First Out) 방식과 관련 효율적 측면에서 비교 분석 한다.

  • PDF

ABRN:An Adaptive Buffer Replacement for On-Demand Multimedia Database Service Systems (ABRN:주문형 멀티미디어 데이터 베이스 서비스 시스템을 위한 버퍼 교체 알고리즘)

  • Jeong, Gwang-Cheol;Park, Ung-Gyu
    • The Transactions of the Korea Information Processing Society
    • /
    • v.3 no.7
    • /
    • pp.1669-1679
    • /
    • 1996
  • In this paper, we address the problem of how to replace huffers in multimedia database systems with time-varying skewed data access. The access pattern in the multimedia database system to support audio-on-demand and video-on-demand services is generally skewed with a few popular objects. In addition the access pattem of the skewed objects has a time-varying property. In such situations, our analysis indicates that conventional LRU(least Recently Used) and LFU(Least Frequently Used) schemes for buffer replacement algorithm(ABRN:Adaptive Buffer Replacement using Neural suited. We propose a new buffer replacement algorithm(ABRN:Adaptive Buffer Replacement using Neural Networks)using a neural network for multimedia database systems with time-varying skewed data access. The major role of our neural network classifies multimedia objects into two classes:a hot set frequently accessed with great popularity and a cold set randomly accessed with low populsrity. For the classification, the inter-arrival time values of sample objects are employed to train the neural network.Our algorithm partitions buffers into two regions to combine the best roperties of LRU and LFU.One region, which contains the 핫셋 objects, is managed by LFU replacement and the other region , which contains the cold set objects , is managed by LRUreplacement.We performed simulation experiments in an actual environment with time-varying skewed data accsee to compare our algorithm to LRU, LFU, and LRU-k which is a variation of LRU. Simulation resuults indicate that our proposed algorthm provides better performance as compared to the other algorithms. Good performance of the neural network-based replacement scheme means that this new approach can be also suited as an alternative to the existing page replacement and prefetching algorithms in virtual memory systems.

  • PDF