• Title/Summary/Keyword: Policy Cache

Search Result 136, Processing Time 0.026 seconds

Delay Reduction by Providing Location Based Services using Hybrid Cache in peer to peer Networks

  • Krishnan, C. Gopala;Rengarajan, A.;Manikandan, R.
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.9 no.6
    • /
    • pp.2078-2094
    • /
    • 2015
  • Now a days, Efficient processing of Broadcast Queries is of critical importance with the ever-increasing deployment and use of mobile technologies. BQs have certain unique characteristics that the traditional spatial query processing in centralized databases does not address. In novel query processing technique, by maintaining high scalability and accuracy, latency is reduced considerably in answering BQs. Novel approach is based on peer-to-peer sharing, which enables us to process queries without delay at a mobile host by using query results cached in its neighboring mobile peers. We design and evaluate cooperative caching techniques to efficiently support data access in ad hoc networks. We first propose two schemes: Cache Data, which caches the data, and Cache Path, which caches the data path. After analyzing the performance of those two schemes, we propose a hybrid approach (Hybrid Cache), which can further improve the performance by taking advantage of Cache Data and Cache Path while avoiding their weaknesses. Cache replacement policies are also studied to further improve the performance. Simulation results show that the proposed schemes can significantly reduce the query delay and message complexity when compared to other caching schemes.

Improving Instruction Cache Performance by Dynamic Management of Cache-Image (캐시 이미지의 동적 관리 방법을 이용한 명령어 캐시 성능 개선)

  • Suh, Hyo-Joong
    • KIISE Transactions on Computing Practices
    • /
    • v.23 no.9
    • /
    • pp.564-571
    • /
    • 2017
  • The burst loading of a pre-created cache-image is an effective method to reduce the instruction cache misses in the early stage of the program execution. It is useful to alleviate the performance degradation as well as the energy inefficiency, which is induced by the concentrated cold misses at the instruction cache. However, there are some defects, including software overhead on the compiler and installer. Furthermore, there are several mismatches as a result of the dynamic properties for specific applications. This paper addresses these issues and proposes a cache-image maintenance/recreation policy that can conduct dynamic management using a hardware-assisted method. The results of the simulation show that the proposed method can maintain the cache-image with a proper size and validity.

Efficient Buffer Allocation Policy for the Adaptive Block Replacement Scheme (적응력있는 블록 교체 기법을 위한 효율적인 버퍼 할당 정책)

  • Choi, Jong-Moo;Cho, Seong-Je;Noh, Sam-Hyuk;Min, Sang-Lyul;Cho, Yoo-Kun
    • Journal of KIISE:Computer Systems and Theory
    • /
    • v.27 no.3
    • /
    • pp.324-336
    • /
    • 2000
  • The paper proposes an efficient buffer management scheme to enhance performance of the disk I/O system. Without any user level information, the proposed scheme automatically detects the block reference patterns of applications by associating block attributes with forward distance of a block. Based on the detected patterns, the scheme applies an appropriate replacement policy to each application. We also present a new block allocation scheme to improve the performance of buffer cache when kernel needs to allocate a cache block due to a cache miss. The allocation scheme analyzes the cache hit ratio of each application based on block reference patterns and allocates a cache block to maximize cache hit ratios of system. These all procedures are performed on-line, as well as automatically at system level. We evaluate the scheme by trace-driven simulation. Experimental results show that our scheme leads to significant improvements in hit ratios of cache blocks compared to the traditional schemes and requires low overhead.

  • PDF

Cache Architecture Design for the Performance Improvement of OpenRISC Core (OpenRISC 코어의 성능향상을 위한 캐쉬 구조 설계)

  • Jung, Hong-Kyun;Ryoo, Kwang-Ki
    • Journal of the Institute of Electronics Engineers of Korea SD
    • /
    • v.46 no.1
    • /
    • pp.68-75
    • /
    • 2009
  • As the recent performance of microprocessor is improving quickly, the necessity of cache is growing because of the increase of the access time of main memory. Every block of direct-mapped cache maps to one cache line. Although the mapping rule is simple, if different blocks map to one cache line, the miss ratio will be higher than the set-associative cache due to conflicts. In this paper, for the improvement of the direct-mapped cache of OpenRISC, 4-way set-associative cache is proposed. Four blocks of the main memory of the proposed cache map to one cache line so that the miss ratio is less than the direct-mapped cache. Pseudo-LRU Policy, which is one of the Line Replacement Policies, is used for decreasing the number of bits that store LRU value. The OpenRISC core including the 4-way set-associative cache was verified with FPGA emulation. As the result of performance measurement using test program, the performance of the OpenRISC core including the 4-way set-associative cache is higher than the previous one by 50% and the decrease of miss ratio is more than 15%.

Study on the Performance Evaluation and Analysis of Mobile Cache Memory

  • Lee, Sangmin;Kim, Jongwan;Kim, Ji Young;Oh, Dukshin
    • Journal of the Korea Society of Computer and Information
    • /
    • v.25 no.6
    • /
    • pp.99-107
    • /
    • 2020
  • In this paper, we analyze the characteristics of mobile cache, which is used to improve the data access speed when executing applications on mobile devices, and verify the importance of mobile cache through a cache data access experiment. The mobile device market has grown at a fast pace over the past decade; however, battery limitations and size, price considerations restrict the usage of fast hardware. Thus, their performance are supplemented by using a memory buffer structure such as the cache memory. The analysis mainly focuses on cache size, hierarchical structure of cache, cache replacement policy, and the effect these features has on mobile performance. For the experimental data, we applied a data set from a microprocessor system study, originally used to test the cache performance. In the experimental results, the average data access speed on a mobile device showed a performance improvement of up to 10 times with the presence of cache memory than without. Accordingly, the cache memory was helpful for the performance improvement of a mobile device when the specifications were identical.

Optimality of Interval Caching Policies in Multimedia Streaming Systems

  • Cho, Kyungwoon;Bahn, Hyokyung
    • International Journal of Internet, Broadcasting and Communication
    • /
    • v.14 no.1
    • /
    • pp.31-36
    • /
    • 2022
  • Interval caching is one of the representative caching strategies used in multimedia streaming systems. However, there has been no theoretical analysis on interval caching. In this paper, we present an optimality proof of the interval caching policy. Specifically, we propose a caching performance model for multimedia streaming systems and show the optimality of the interval caching policy based on this model.

A Study of File Replacement Policy in Data Grid Environments (데이터 그리드 환경에서 파일 교체 정책 연구)

  • Park, Hong-Jin
    • The KIPS Transactions:PartA
    • /
    • v.13A no.6 s.103
    • /
    • pp.511-516
    • /
    • 2006
  • The data grid computing provides geographically distributed storage resources to solve computational problems with large-scale data. Unlike cache replacement policies in virtual memory or web-caching replacement, an optimal file replacement policy for data grids is the one of the important problems by the fact that file size is very large. The traditional file replacement policies such as LRU(Least Recently Used) LCB-K(Least Cost Beneficial based on K), EBR(Economic-based cache replacement), LVCT(Least Value-based on Caching Time) have the problem that they have to predict requests or need additional resources to file replacement. To solve theses problems, this paper propose SBR-k(Sized-based replacement-k) that replaces files based on file size. The results of the simulation show that the proposed policy performs better than traditional policies.

An Adaptive Cache Replacement Policy for Web Proxy Servers (웹 프락시 서버를 위한 적응형 캐시 교체 정책)

  • Choi, Seung-Lak;Kim, Mi-Young;Park, Chang-Sup;Cho, Dae-Hyun;Lee, Yoon-Joon
    • Journal of KIISE:Computer Systems and Theory
    • /
    • v.29 no.6
    • /
    • pp.346-353
    • /
    • 2002
  • The explosive increase of World Wide Web usage has incurred significant amount of network traffic and server load. To overcome these problems, web proxy caching replicates frequently requested documents in the web proxy closer to the users. Cache utilization depends on the replacement policy which tries to store frequently requested documents in near future. Temporal locality and Zipf frequency distribution, which are commonly observed in web proxy workloads, are considered as the important properties to predict the popularity of documents. In this paper, we propose a novel cache replacement policy, called Adaptive LFU (ALFU), which incorporates 1) Zipf frequency distribution by utilizing LFU and 2) temporal locality adaptively by measuring the amount of the popularity reduction of documents as time passed efficiently. We evaluate the performance of ALFU by comparing it to other policies via trace-driven simulation. Experimental results show that ALFU outperforms other policies.

Cache Policy based on Producer Distance to Reduce Response Time in CCN (CCN에서 응답시간 감소를 위한 생산자 거리 기반 캐시정책)

  • Kim, Keon;Kwon, Tae-Wook
    • The Journal of the Korea institute of electronic communication sciences
    • /
    • v.16 no.6
    • /
    • pp.1121-1132
    • /
    • 2021
  • Nowadays, it is more difficult to find people who do not use mobile devices such as smartphones and tablets. Contents that can be accessed at the touch of a finger is overflowing. However, the existing network has a structure in which it is difficult to efficiently respond to the problems caused by overflowing contents. In particular, the bottleneck problem that occurs when multiple users intensively request content from the server at the same time is a representative problem. To solve this problem, the CCN has emerged as an alternative to future networks. CCN uses the network bandwidth efficiently through the In-Network Cache function of the intermediate node to improve the traffic required for user to request to reach the server, to reduce response time, and to distribute traffic concentration within the network. I propose a cache policy that can improve efficiency in such a CCN environment.

A Study of Weighted Disk Cache Method for World Wide Web (WWW를 위한 가중화 디스크 캐시 기법에 대한 연구)

  • 박해우;강병욱
    • Proceedings of the IEEK Conference
    • /
    • 2002.06c
    • /
    • pp.153-156
    • /
    • 2002
  • As the use of world wide web is increasing, the number of connections to servers is increasing also. These interactions increase the load of networks and servers. therefore efficient caching strategies for web documents are needed to reduce server load and network traffics by migrating copies of server files closer to the clients that use those files. As One idea of caching policy, we propose a Weighted Disk Cache Replacement Policy(WDCRP) which analyses user's interaction to WWW and adds weight value to each web document. Especially the WDCRP takes account of the history data of cache log, the characteristics of Web requests and the importance of user interactive-actions.

  • PDF