• Title/Summary/Keyword: Cache server

Search Result 220, Processing Time 0.033 seconds

Exploitation of Multi-Versions based on Callback Locking in a Client-Server DBMS Environment (클라이언트-서버 DBMS 환경에서 콜백 잠금 기반 다중 버전의 활용)

  • 강흠근;민준기;전석주;정진완
    • Journal of KIISE:Databases
    • /
    • v.31 no.5
    • /
    • pp.457-467
    • /
    • 2004
  • The efficiency of algorithms managing data caches has a major impact on the performance of systems that utilize client-side data caching. In these systems, two versions of data can be maintained without an additional space overhead of the server by exploiting the replication of data in the server's buffer and clients' caches. In this paper, we present a new cache consistency algorithm employing versions: Two Versions-Callback Locking (2V-CBL). Our experimental results indicate that 2V-CBL provides good performance, particularly outperforms a leading cache consistency algorithm, Asynchronous Avoidance-based Gache Consistency, when some clients run only read-only transactions.

Replacement Algorithm Selection Mechanism Considering File Size for Web Cache Server

  • Sontisiri, Tanasun;Sopechoke, Pawin;Thipchaksurat, Sakchai;Varakulsiripunth, Ruttikorn
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 2004.08a
    • /
    • pp.1084-1089
    • /
    • 2004
  • This paper describes the improvement of web cache server by scoping in replacement algorithm of data which are collected from the clients. We have found that each replacement algorithm is suitable for each type of data in the web pages. Therefore, we introduce the mechanism to select the replacement algorithm depending on the size of data called the Replacement Algorithm Selection Mechanism (RASM). RASM allows the web cache server to have the suitable replacement algorithm for each type of data. As the result, the byte hit ratio of web cache server can be increased and the congestion in the network can be alleviated.

  • PDF

A New Hybrid Architecture for Cooperative Web Caching

  • Baek, Jin-Suk;Kaur, Gurpreet;Yang, Jung-Hoon
    • Journal of Ubiquitous Convergence Technology
    • /
    • v.2 no.1
    • /
    • pp.1-11
    • /
    • 2008
  • An effective solution to the problems caused by the explosive growth of World Wide Web is a web caching that employing an additional server, called proxy cache, between the clients and main server for caching the popular web objects near the clients. However, a single proxy cache can easily become the bottleneck. Deploying groups of cooperative caches provides scalability and robustness by eliminating the limitations caused by a single proxy cache. Two common architectures to implement the cooperative caching are hierarchical and distributed caching systems. Unfortunately, both architectures suffer from performance limitations. We propose an efficient hybrid caching architecture eliminating these limitations by using both the hierarchical and same level caches. Our performance evaluation with our investigated simulator shows that the proposed architecture offers the best of both existing architectures in terms of cache hit rate, the number of query messages from clients, and response time.

  • PDF

Performance Impact of Large File Transfer on Web Proxy Caching: A Case Study in a High Bandwidth Campus Network Environment

  • Kim, Hyun-Chul;Lee, Dong-Man;Chon, Kil-Nam;Jang, Beak-Cheol;Kwon, Tae-Kyoung;Choi, Yang-Hee
    • Journal of Communications and Networks
    • /
    • v.12 no.1
    • /
    • pp.52-66
    • /
    • 2010
  • Since large objects consume substantial resources, web proxy caching incurs a fundamental trade-off between performance (i.e., hit-ratio and latency) and overhead (i.e., resource usage), in terms of caching and relaying large objects to users. This paper investigates how and to what extent the current dedicated-server based web proxy caching scheme is affected by large file transfers in a high bandwidth campus network environment. We use a series of trace-based performance analyses and profiling of various resource components in our experimental squid proxy cache server. Large file transfers often overwhelm our cache server. This causes a bottleneck in a web network, by saturating the network bandwidth of the cache server. Due to the requests for large objects, response times required for delivery of concurrently requested small objects increase, by a factor as high as a few million, in the worst cases. We argue that this cache bandwidth bottleneck problem is due to the fundamental limitations of the current centralized web proxy caching model that scales poorly when there are a limited amount of dedicated resources. This is a serious threat to the viability of the current web proxy caching model, particularly in a high bandwidth access network, since it leads to sporadic disconnections of the downstream access network from the global web network. We propose a peer-to-peer cooperative web caching scheme to address the cache bandwidth bottleneck problem. We show that it performs the task of caching and delivery of large objects in an efficient and cost-effective manner, without generating significant overheads for participating peers.

A Weighted-window based Cache Replacement Policy for Streaming Cache Server (스트리밍 캐쉬 서버를 위한 가중치 윈도우 기반의 캐쉬 교체 정책)

  • 오재학;차호정;박병준
    • Journal of KIISE:Computer Systems and Theory
    • /
    • v.30 no.10
    • /
    • pp.556-568
    • /
    • 2003
  • This paper presents and analyzes the performance of a weighted-window based cache replacement policy for the efficient media caching in streaming media cache servers. The proposed policy makes, for each cached object, use of the reference count, reference time, amount of media delivered to clients and, in particular, the periodic patterns of user requests. Also, by giving weights to the recently referenced media contents, the replacement policy adequately and swiftly reflects the ever-changing characteristics of users preferences. The simulation studies show that the performance of the proposed policy has improved over the conventional policies such as LRU, LFU and SEG - in terms of hit ratio, byte hit ratio, delayed start and cache input.

Location tracking using cache in distributed call processing architecture for PCS (PCS를 위한 분산 호 처리 구조(DCPA)에서 캐쉬를 이용한 위치 추적)

  • 박선영;이원열;한기준
    • Journal of the Korean Institute of Telematics and Electronics S
    • /
    • v.34S no.8
    • /
    • pp.27-36
    • /
    • 1997
  • In this paper, we propose a locating strategy using a cache based on the distributed call processing stucture. In our stategy, each originating call server, withc is responsible for call setup, has a cache to memorize the service ratio of call server which serves the called terminal.We present performance comparison of the proposed strategy with the existing strategy via simulation. We can reduce the cost for call setup because the originating call server can simpley establish a call without tracing the location of the called terminal. The proposed strategy also reduceds network loads caused by signaling at call setup time and lessens the frequency of queries for the database systems which can be a bottleneck when it serves many subscriber.

  • PDF

Web Proxy Cache Replacement Algorithms using Object Type Partition (개체 타입별 분할공간을 이용한 웹 프락시 캐시의 대체 알고리즘)

  • Soo-haeng, Lee;Sang-bang, Choi
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.27 no.5C
    • /
    • pp.399-410
    • /
    • 2002
  • Web cache, which is functionally another word of proxy server, is located between client and server. Web cache has a limited storage area although it has broad bandwidth between client and proxy server, which are usually connected through LAN. Because of limited storage capacity, existing objects in web cache can be deleted for new objects by some rules called replacement algorithm. Hit rate and byte-hit rate are general metrics to evaluate replacement algorithms. Most of the replacement algorithms do satisfy only one metric, or sometimes none of them. In this paper, we propose two replacement algorithms to achieve both high hit rate and byte-hit rate with great satisfaction. In the first algorithm, the cache is appropriately partitioned according to file types as a basic model. In the second algorithm, the cache is composed of two levels; the upper level cache is managed by the basic algorithm, but the lower level is collectively used for all types of files as a shared area. To show the performance of the proposed algorithms, we evaluate hit rate and byte-hit rate of the proposed replacement algorithms using the trace driven simulation.

Selective Cache Consistency Scheme to Enlarge Autonomy of Mobile Host in Mobile Computing Environments (이동 컴퓨팅 환경에서 이동 호스트의 자치성 증대를 위한 선택적 캐쉬 일관성 유지 기법)

  • Kim, Hee-Sook;Hwang, Byung-Yeon
    • The KIPS Transactions:PartD
    • /
    • v.10D no.4
    • /
    • pp.655-660
    • /
    • 2003
  • The cache used by mobile host is an important device that recovers the weak points of limited power and bandwidth, in mobile computing environments. However, it has to stand and maintain the consistency with the server data. In this paper, we propose a 'Selective Cache Consistency Scheme'. The server allows an effective broadcasting by selecting data of high usability using 'Cache State Table' and 'Data Access Table'. Moreover, this scheme prevents the loss of data that nay occur by a long period of disconnection, by asynchronous broadcasting and transmitting those broadcast data preserved in the server. This also allows user to possess the latest data. Through experiments, we have found that the enlargement of autonomy is possible by reducing the dependence of server.

Implementation of Client Authentication System on Transparency Cache (투명 캐시에서의 사용자 인증 시스템 구현)

  • Kim, Seong-Rak;Gu, Yong-Wan
    • The KIPS Transactions:PartC
    • /
    • v.9C no.2
    • /
    • pp.181-188
    • /
    • 2002
  • There are recently a lot of inconvenience because every client should be set to the proxy server on the browser in order to control the access by means of the client authentication in the proxy server. The client authentication technology using the transparency cache in this paper will be transparently used for every user in the internet which option of the authentication function is simply set in the cache server. In addition, the administrator will get the benefit since he can control the traffic of each client and strengthen the security. And also, this system is expected to use in the eCRM deeply rotated to the tendency of the client in the field of the e-commerce like shopping mall in the internet since the administrator can monitor the pattern of the client using the internet. This technique can be applied to the company affiliated research center, the EC website, and the military where it is essential for the tight security even though there are no additional security devices.

Asynchronous Cache Consistency Technique (비동기적 캐쉬 일관성 유지 기법)

  • 이찬섭
    • Journal of the Korea Society of Computer and Information
    • /
    • v.9 no.2
    • /
    • pp.33-40
    • /
    • 2004
  • According as client/server is generalized by development of computer performance and information communication technology, Servers uses local cache for extensibility and early response time, and reduction of limited bandwidth. Consistency of cached data need between server and client this time and much technique are proposed according to this. This Paper improved update frequency cache consistency in old. Existent consistency techniques is disadvantage that response time is late because synchronous declaration or abort step increases because delaying write intention declaration. Techniques that is proposed in this paper did to perform referring update time about object that page request or when complete update operation happens to solve these problem. Therefore, have advantage that response is fast because could run write intention declaration or update by sel_mode electively asynchronously when update operation consists and abort step decreases and clearer selection.

  • PDF