• Title/Summary/Keyword: Cache Replacement Method

Search Result 26, Processing Time 0.022 seconds

A Cache Policy Based on Producer Popularity-Distance in CCN (CCN에서 생성자 인기도 및 거리 기반의 캐시정책)

  • Min, Ji-Hwan;Kwon, Tae-Wook
    • The Journal of the Korea institute of electronic communication sciences
    • /
    • v.17 no.5
    • /
    • pp.791-800
    • /
    • 2022
  • CCN, which has emerged to overcome the limitations of existing network structures, enables efficient networking by changing the IP Address-based network method to the Content-based network method. At this time, the contents are stored on each node(router) rather than on the top server, and considering the limitation of storage capacity, it is very important to determine which contents to store and release through the cache policy, and there are several cache policies that have been studied so far. In this paper, two of the existing cache policies, producer popularity-based and distance-based, were mixed. In addition, the mixing ratio was set differently to experiment, and we proved which experiement was the most efficient one.

Caching and Prefetching Policies Using Program Page Reference Patterns on a File System Layer for NAND Flash Memory (NAND 플래시 메모리용 파일 시스템 계층에서 프로그램의 페이지 참조 패턴을 고려한 캐싱 및 선반입 정책)

  • Kim, Gyeong-San;Kim, Seong-Jo
    • Proceedings of the IEEK Conference
    • /
    • 2006.06a
    • /
    • pp.777-778
    • /
    • 2006
  • In this thesis, we design and implement a Flash Cache Core Module (FCCM) which operates on the YAFFS NAND flash memory. The FCCM applies memory replacement policy and prefetching policy based on the page reference pattern of applications. Also, implement the Clean-First memory replacement technique considering the characteristics of flash memory. In this method the decision is made according to page hit to apply prefetched waiting area. The FCCM decrease I/O hit frequency up to 37%, Compared with the linux cache and prefetching policy. Also, it operated using less memory for prefetching(maximum 24% and average 16%) compared with the linux kernel.

  • PDF

Fault Tolerant Cache for Soft Error (소프트에러 결함 허용 캐쉬)

  • Lee, Jong-Ho;Cho, Jun-Dong;Pyo, Jung-Yul;Park, Gi-Ho
    • The Transactions of The Korean Institute of Electrical Engineers
    • /
    • v.57 no.1
    • /
    • pp.128-136
    • /
    • 2008
  • In this paper, we propose a new cache structure for effective error correction of soft error. We added check bit and SEEB(soft error evaluation block) to evaluate the status of cache line. The SEEB stores result of parity check into the two-bit shit register and set the check bit to '1' when parity check fails twice in the same cache line. In this case the line where parity check fails twice is treated as a vulnerable to soft error. When the data is filled into the cache, the new replacement algorithm is suggested that it can only use the valid block determined by SEEB. This structure prohibits the vulnerable line from being used and contributes to efficient use of cache by the reuse of line where parity check fails only once can be reused. We tried to minimize the side effect of the proposed cache and the experimental results, using SPEC2000 benchmark, showed 3% degradation in hit rate, 15% timing overhead because of parity logic and 2.7% area overhead. But it can be considered as trivial for SEEB because almost tolerant design inevitably adopt this parity method even if there are some overhead. And if only parity logic is used then it can have $5%{\sim}10%$ advantage than ECC logic. By using this proposed cache, the system will be protected from the threat of soft error in cache and the hit rate can be maintained to the level without soft error in the cache.

Efficient Cooperative Caching Algorithm for Distributed File Systems (분산 파일시스템을 위한 효율적인 협력캐쉬 알고리즘)

  • 박새미;이석재;유재수
    • Proceedings of the Korea Contents Association Conference
    • /
    • 2003.11a
    • /
    • pp.234-244
    • /
    • 2003
  • In distributed file-systems, cooperative caching algorithm which owns the data cached at each node jointly is used to reduce an expense of disk access. Cooperative caching algorithm is the method that increases a cache hit-ratio and decrease a disk access as it holds the cache information of distributed systems in common and makes cache larger virtually. Recently, several cooperative caching algorithms decrease the message costs by using approximate information of the cache and increase the cache hit-ratio by using local and global cache fields dynamically. And they have an advantage that increases the whole field hit-ratio by sending a replaced block to the idel node on cache replacement in order to maintain the replaced block in the cache field. However the wrong approximate information deteriorates the performance, the concistency maintenance goes to great expense to exchange messeges and the cost that manages Age-information of each node to choose the idle node increases. In this thesis, we propose a cooperative cache algorithm that maintains correct cache information, minimizes the maintance cost for consistency and the management cost for cache Age-information. Also, we show the superiority of our algorithm through the performance evaluation.

  • PDF

Wireless Caching Algorithm Based on User's Context in Smallcell Environments (소형셀 환경에서 사용자 컨텍스트 기반 무선 캐시 알고리즘)

  • Jung, Hyun Ki;Jung, Soyi;Lee, Dong Hak;Lee, Seung Que;Kim, Jae-Hyun
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.41 no.7
    • /
    • pp.789-798
    • /
    • 2016
  • In this paper, we propose a cache algorithm based on user's context for enterprise/urban smallcell environments. The smallcell caching method is to store mobile users' data traffic at a storage which is equipped in smallcell base station and it has an effect of reducing core networks traffic volume. In our algorithm, contrary to existing smallcell cache algorithms, the cache storage is equipped in a edge server by using a concept of the Mobile Edge Computing. In order to reflect user's characteristics, the edge server classifies users into several groups based on user's context. Also the edge server changes the storage size and the cache replacement frequency of each group to improve the cache efficiency. As the result of performance evaluation, the proposed algorithm can improve the cache hit ratio by about 11% and cache efficiency by about 5.5% compared to the existing cache algorithm.

Reducing Outgoing Traffic of Proxy Cache by Using Client-Cluster

  • Kim Kyung-Baek;Park Dae-Yeon
    • Journal of Communications and Networks
    • /
    • v.8 no.3
    • /
    • pp.330-338
    • /
    • 2006
  • Many web cache systems and policies concerning them have been proposed. These studies, however, consider large objects less useful than small objects in terms of performance, and evict them as soon as possible. Even if this approach increases the hit rate, the byte hit rate decreases and the connections occurring over congested links to outside networks waste more bandwidth in obtaining large objects. This paper puts forth a client-cluster approach for improving the web cache system. The client-cluster is composed of the residual resources of clients and utilizes them as exclusive storage for large objects. This proposed system achieves not only a high hit rate but also a high byte hit rate, while reducing outgoing traffic. The distributed hash table (DHT) based peer-to-peer lookup protocol is utilized to manage the client-cluster. With the natural characteristics of this protocol, the proposed system with the client-cluster is self-organizing, fault-tolerant, well-balanced, and scalable. Additionally, the large objects are managed with an index based allocation method, which balances the loads of all clients well. The performance of the cache system is examined via a trace driven simulation and an effective enhancement of the proxy cache performance is demonstrated.

A novel page replacement policy associated with ACT-R inspired by human memory retrieval process (인간 기억 인출 과정을 응용하여 설계된 ACT-R 기반 페이지 교체 정책)

  • Roh, Hong-Chan;Park, Sang-Hyun
    • The KIPS Transactions:PartD
    • /
    • v.18D no.1
    • /
    • pp.1-8
    • /
    • 2011
  • The cache structure, which is designed for assuring fast accesses to frequently accessed data, resides on the various levels of computer system hierarchies. Many studies on this cache structure have been conducted and thus many page-replacement algorithms have been proposed. Most of page-replacement algorithms are designed on the basis of heuristic methods by using their own criteria such as how recently pages are accessed and how often they are accessed. This data-retrieval process in computer systems is analogous to human memory retrieval process since the retrieval process of human memory depends on frequency and recency of the retrieval events as well. A recent study regarding human memory cognition revealed that the possibility of the retrieval success and the retrieval latency have a strong correlation with the frequency and recency of the previous retrieval events. In this paper, we propose a novel page-replacement algorithm by utilizing the knowledge from the recent research regarding human memory cognition. Through a set of experiments, we demonstrated that our new method presents better hit-ratio than the LRFU algorithm which has been known as the best performing page-replacement algorithm for DBMS caches.

A Study of Weighted Disk Cache Method for World Wide Web (WWW를 위한 가중화 디스크 캐시 기법에 대한 연구)

  • 박해우;강병욱
    • Proceedings of the IEEK Conference
    • /
    • 2002.06c
    • /
    • pp.153-156
    • /
    • 2002
  • As the use of world wide web is increasing, the number of connections to servers is increasing also. These interactions increase the load of networks and servers. therefore efficient caching strategies for web documents are needed to reduce server load and network traffics by migrating copies of server files closer to the clients that use those files. As One idea of caching policy, we propose a Weighted Disk Cache Replacement Policy(WDCRP) which analyses user's interaction to WWW and adds weight value to each web document. Especially the WDCRP takes account of the history data of cache log, the characteristics of Web requests and the importance of user interactive-actions.

  • PDF

PECAN: Peer Cache Adaptation for Peer-to-Peer Video-on-Demand Streaming

  • Kim, Jong-Tack;Bahk, Sae-Woong
    • Journal of Communications and Networks
    • /
    • v.14 no.3
    • /
    • pp.286-295
    • /
    • 2012
  • To meet the increased demand of video-on-demand (VoD) services, peer-to-peer (P2P) mesh-based multiple video approaches have been recently proposed, where each peer is able to find a video segment interested without resort to the video server. However, they have not considered the constraint of the server's upload bandwidth and the fairness between upload and download amounts at each peer. In this paper, we propose a novel P2P VoD streaming system, named peer cache adaptation (PECAN) where each peer adjusts its cache capacity adaptively to meet the server's upload bandwidth constraint and achieve the fairness. For doing so, we first propose a new cache replacement algorithm that designs the number of caches for a segment to be proportional to its popularity. Second, we mathematically prove that if the cache capacity of a peer is proportional to its segment request rate, the fairness between upload and download amounts at each peer can be achieved. Third, we propose a method that determines each peer's cache capacity adaptively according to the constraint of the server's upload bandwidth. Against the proposed design objective, some selfish peers may not follow our protocol to increase their payoff. To detect such peers, we design a simple distributed reputation and monitoring system. Through simulations, we show that PECAN meets the server upload bandwidth constraint, and achieves the fairness well at each peer. We finally verify that the control overhead in PECAN caused by the search, reputation, and monitoring systems is very small, which is an important factor for real deployment.

Instructions and Data Prefetch Mechanism using Displacement History Buffer (변위 히스토리 버퍼를 이용한 명령어 및 데이터 프리페치 기법)

  • Jeong, Yong Su;Kim, JinHyuk;Cho, Tae Hwan;Choi, SangBang
    • Journal of the Institute of Electronics and Information Engineers
    • /
    • v.52 no.10
    • /
    • pp.82-94
    • /
    • 2015
  • In this paper, we propose hardware prefetch mechanism with an efficient cache replacement policy by giving priority to the trigger block in which a spatial region and producing a spatial region by using the displacement field. It could be taken into account the sequence of the program since a history is based on the trigger block of history record, and it could be quickly prefetching the instructions or data address by adding a stored value to the trigger address and displacement field since a history is stored as a displacement value. Also, we proposed a method of replacing at random by the cache replacement policy from the low priority block when the cache area is full after giving priority to the trigger block. We analyzed using the memory simulator program gem5 and PARSEC benchmark to assess the performance of the hardware prefetcher. As a result, compared to the existing hardware prefecture to generate the spatial region using a bit vector, L1 data cache miss rate was reduced about 44.5% on average and an average of 26.1% of L1 instruction misses occur. In addition, IPC (Instruction Per Cycle) showed an improvement of about 23.7% on average.