• Title/Summary/Keyword: Byte Hit-Ratio

Search Result 9, Processing Time 0.027 seconds

Cache Policies for WWW Sewers to Reduce Workload (웹 서버 작업부하 감소를 위한 캐시 정책)

  • Lim Jaehyun
    • Journal of Internet Computing and Services
    • /
    • v.5 no.6
    • /
    • pp.103-110
    • /
    • 2004
  • In this paper we study and analyze the influence of caching strategies on the performance of WWW servers, We propose a new strategy called file type based caching that aims to obtain a well-balanced mixture between large and small files in the cache, and moreover, it provides good performance for both small and large file as expected, By using the type based caching good results are obtained for both the hit rate and the byte hit rate.

  • PDF

Web Caching using File Type (파일 타입을 이용한 웹 캐싱)

  • Lim, Jae-Hyun;Lee, Jun-Yeon
    • The KIPS Transactions:PartC
    • /
    • v.9C no.6
    • /
    • pp.961-968
    • /
    • 2002
  • This paper proposes a new access method which is to considered the high variability in World Wide Web and manage the web cache space. Instead of using a single cache, we divide a cache and store all documents according to their file types. Proposed method was compares with current cache management policies using LFU, LRU and SIZE base algorithm. Using two different workload, we show the improvement hitting ratio and byte hitting ratio through simulating on the file type caching.

Replacement Algorithm Selection Mechanism Considering File Size for Web Cache Server

  • Sontisiri, Tanasun;Sopechoke, Pawin;Thipchaksurat, Sakchai;Varakulsiripunth, Ruttikorn
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 2004.08a
    • /
    • pp.1084-1089
    • /
    • 2004
  • This paper describes the improvement of web cache server by scoping in replacement algorithm of data which are collected from the clients. We have found that each replacement algorithm is suitable for each type of data in the web pages. Therefore, we introduce the mechanism to select the replacement algorithm depending on the size of data called the Replacement Algorithm Selection Mechanism (RASM). RASM allows the web cache server to have the suitable replacement algorithm for each type of data. As the result, the byte hit ratio of web cache server can be increased and the congestion in the network can be alleviated.

  • PDF

Design and Performance Analysis of an Efficient Cache Managing Strategy for Web Services (웹 서비스를 위한 효율적인 캐쉬 관리 전략의 설계 및 성능 평가)

  • Moon, Jin-Yong
    • Journal of Digital Contents Society
    • /
    • v.9 no.4
    • /
    • pp.653-659
    • /
    • 2008
  • With the recent explosion in using of the Internet, the problem of caching web objects has gained considerable importance. Caching on the Internet differs from traditional caching in several ways. Especially, the conventional caching algorithms are not well suited for the Internet caching. The poor performance is mainly due to its unfair treatment of small objects since all the objects are treated the same even though they differ in size. In this paper, I give an overview of caching policies designed for web objects, and provide a new algorithm of my own. I also have performed trace-driven simulations about variable-size replacement algorithms, and derived a new algorithm to improve byte hit-ratio by classifying objects based on their sizes.

  • PDF

A Web Cache Replacement Policy in Consideration of Netwok Response Time (네트워크 응답시간을 고려한 웹 캐시 교체 정책)

  • 신은희;서진모;신승훈;박승규
    • Proceedings of the Korean Information Science Society Conference
    • /
    • 2003.04a
    • /
    • pp.133-135
    • /
    • 2003
  • 최근 인터넷 기반 서비스 환경의 질적개선과 이에 따른 이용자 증가에 따라. 오디오나 동영상과 같은 규모가 크고 동적 특성을 가진 웹 컨텐츠의 수가 증가하는 추세이다. 이에 따라 'hit ratio' 뿐만 아니라 오브젝트의 크기를 기반으로 하는 'byte hit ratio' 또한 캐시 성능을 평가하는 중요한 측정 요소가 되었다. 이러한 측정기준을 대상으로 할 때, 기존의 웹 캐시 교체 정책중 HP연구소의 GDSF (Greedy-Dual-Size with Frequency)알고리즘과 LFU-DA(LFU with Dynamic Aging)알고리즘이 우수한 성능을 보이는 것으로 알려져있는데, 이러한 기존의 웹 캐시 교체 정책은 서버와의 네트워크 상태를 고려하지 않은 정책이고, 이에 따라 네트워크 상태에 따른 전송비용의 차이를 반영하지 못하고 있다. 따라서 본 논문에서는 서버와 웹 캐시 간의 네트워크 상태를 반영할 수 있는 캐시 교체 정책을 제안하고, 이에 대한 실험을 수행하였으며, 그 결과 사용자의 요구에 대한 응답시간의 감소 효과를 얻을 수 있었다.

  • PDF

The Effects of Cache Memory on the System Bus Traffic (캐쉬 메모리가 버스 트래픽에 끼치는 영향)

  • 조용훈;김정선
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.21 no.1
    • /
    • pp.224-240
    • /
    • 1996
  • It is common sense for at least one or more levels of cache memory to be used in these day's computer systems. In this paper, the impact of the internal cache memory organization on the performance of the computer is investigated by using a simulator program, which is wirtten by authors and run on SUN SPARC workstation, with several real execution, with several real execution trace files. 280 cache organizations have been simulated using n-way set associative mapping and LRU(Least Recently Used) replacement algorithm with write allocation policy. As a result, 16-way setassociative cache is the best configuration, and when we select 256KB cache memory and 64 byte line size, the bus traffic ratio was decreased compared to that of the noncache system so that a single bus could support almost 7 processors without any delay and degradationof high ratio(hit ratio was 99.21%). The smaller the line size we choose, the little lower hit ratio we can get, but the more processors can be supported by a single bus(maximum 18 processors). Therefore, using a proper cache memory organization can make a single bus structure be able to support multiple processors without any performance degradation.

  • PDF

A Weighted-window based Cache Replacement Policy for Streaming Cache Server (스트리밍 캐쉬 서버를 위한 가중치 윈도우 기반의 캐쉬 교체 정책)

  • 오재학;차호정;박병준
    • Journal of KIISE:Computer Systems and Theory
    • /
    • v.30 no.10
    • /
    • pp.556-568
    • /
    • 2003
  • This paper presents and analyzes the performance of a weighted-window based cache replacement policy for the efficient media caching in streaming media cache servers. The proposed policy makes, for each cached object, use of the reference count, reference time, amount of media delivered to clients and, in particular, the periodic patterns of user requests. Also, by giving weights to the recently referenced media contents, the replacement policy adequately and swiftly reflects the ever-changing characteristics of users preferences. The simulation studies show that the performance of the proposed policy has improved over the conventional policies such as LRU, LFU and SEG - in terms of hit ratio, byte hit ratio, delayed start and cache input.

Cache Replacement and Coherence Policies Depending on Data Significance in Mobile Computing Environments (모바일 컴퓨팅 환경에서 데이터의 중요도에 기반한 캐시 교체와 일관성 유지)

  • Kim, Sam-Geun;Kim, Hyung-Ho;Ahn, Jae-Geun
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.36 no.2A
    • /
    • pp.149-159
    • /
    • 2011
  • Recently, mobile computing environments are becoming rapidly common. This trend emphasizes the necessity of accessing database systems on fixed networks from mobile platforms via wireless networks. However, it is not an appropriate way that applies the database access methods for traditional computing environments to mobile computing environments because of their essential restrictions. This paper suggests a new agent-based mobile database access model and also two functions calculating data significance scores to choose suitable data items for cache replacement and coherence policies. These functions synthetically reflect access term, access frequency and tendency, update frequency and tendency, and data item size distribution. As the result of simulation experiment, our policies outperform LRU, LIX, and SAIU policies in aspects of decrement of access latency, improvement of cache byte hit ratio, and decrease of cache byte pollution ratio.

Segment-based Cache Replacement Policy in Transcoding Proxy (트랜스코딩 프록시에서 세그먼트 기반 캐쉬 교체 정책)

  • Park, Yoo-Hyun;Kim, Hag-Young;Kim, Kyong-Sok
    • The KIPS Transactions:PartA
    • /
    • v.15A no.1
    • /
    • pp.53-60
    • /
    • 2008
  • Streaming media has contributed to a significant amount of today's Internet Traffic. Like traditional web objects, rich media objects can benefit from proxy caching, but caching streaming media is more of challenging than caching simple web objects, because the streaming media have features such as huge size and high bandwidth. And to support various bandwidth requirements for the heterogeneous ubiquitous devices, a transcoding proxy is usually necessary to provide not only adapting multimedia streams to the client by transcoding, but also caching them for later use. The traditional proxy considers only a single version of the objects, whether they are to be cached or not. However the transcoding proxy has to evaluate the aggregate effect from caching multiple versions of the same object to determine an optimal set of cache objects. And recent researches about multimedia caching frequently store initial parts of videos on the proxy to reduce playback latency and archive better performance. Also lots of researches manage the contents with segments for efficient storage management. In this paper, we define the 9-events of transcoding proxy using 4-atomic events. According to these events, the transcoding proxy can define the next actions. Then, we also propose the segment-based caching policy for the transcoding proxy system. The performance results show that the proposing policy have a low delayed start time, high byte-hit ratio and less transcoding data.