• Title/Summary/Keyword: Caching Algorithm

Search Result 105, Processing Time 0.025 seconds

Mobility-Aware Edge Caching Algorithm with Dynamic Content Popularity (동적 인기도 콘텐츠를 활용한 이동성 인식 엣지 캐싱 알고리즘)

  • Lee, Tae-Yoon;Lee, SuKyoung
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2022.11a
    • /
    • pp.135-137
    • /
    • 2022
  • 이동성 기반의 기존 엣지 캐싱 연구에서는 인기도가 짧은 시간 급격하게 변화하는 SNM(Shot Noise Model) 콘텐츠를 반영하지 않았다. 동적 인기도 특성을 다루지 않는 경우, 잦은 캐시 미스가 발생하므로 SNM 콘텐츠를 고려하는 것은 중요하다. 이에 본 논문은 이동성을 고려한 기존 연구에 SNM 콘텐츠를 함께 고려하고, 시뮬레이션을 통해 기존 연구 대비 제안 알고리즘의 향상된 캐시 적중률을 확인한다.

ACASH: An Adaptive Web Caching Method with Heterogeneity of Web Object and Reference Characteristics (ACASH: 웹 객체의 이질성과 참조특성 기반의 적응형 웹 캐싱 기법)

  • 고일석;임춘성;나윤지
    • Journal of KIISE:Information Networking
    • /
    • v.31 no.3
    • /
    • pp.305-313
    • /
    • 2004
  • The use of a cache for a storing and processing of Web object is becoming larger. Also, many studies for efficient management of storing scope on cache are performed actively. Web caching technique have many differences with traditional techniques. Particularly, a heterogeneity of Web object which is a processing unit of Web caching and a variation of Web object reference characteristic with time are the important causes to decrease performance of existing techniques. In this study, We proposed the ACASH which was new web caching technique. As ACASH divided and managed Web object and a cache scope with a heterogeneity, It can reduced a heterogeneity variation of an object. Also, it is reflecting a variation of object reference characteristics with time adaptively. In the experiment, We verified that the performance of ACASH was improved than existing techniques on the two experiment model which considered a heterogeneity of an object.

VOD Server using Web-Caching in Head-End-Network (Head-End-Network에서 Web-Caching을 사용한 VOD 서버)

  • Kim Backhyun;Hwang Taejune;Kim Iksoo
    • Journal of Internet Computing and Services
    • /
    • v.5 no.1
    • /
    • pp.15-23
    • /
    • 2004
  • This paper presents distributive web-caching technique on the Head-End-Network (HNET) to solve excessive load of multimedia sever and inefficient use of network resources, The proposed web-caching is not concentration of load for a specific Head-End-Node(HEN), and it has advantage that there is no copy of a specific item on distributive HENs. This technique distributively stores on HENs partial streams of requested videos from clients connected to HENs and the order of store streams follows the order of request of identical video items from HENs. Thus, storing streams on each HEN that requests sevice are different. When a client requests the cached video on some HENs, the HEN that connects him receives streams in the order of store from HENs which stored them and it transmits them to him. These procedures are performed under the control of SA. This technique uses cache replacement algorithm that the combination of LFU, LRU and the last remove for the first stream of videos In order to reduce end-to-end delay.

  • PDF

A efficient caching scheme of Continuous stream service on the internet (인터넷 상에서 비디오 스트림을 위한 효과적인 캐슁기법)

  • 김종경;남현우
    • Journal of the Korea Computer Industry Society
    • /
    • v.4 no.12
    • /
    • pp.1083-1092
    • /
    • 2003
  • This paper propose method to minimize service delay time based in popularity when service video stream in internet Result that measure performance of algorithm that propose considering correlation of cache space all of access frequency and each servers showed performance elevation of degree of triple than elevation of cache practical use rate 15% and interval caching about service laytency than agent caching scheme.

  • PDF

Web Caching using File Type (파일 타입을 이용한 웹 캐싱)

  • Lim, Jae-Hyun;Lee, Jun-Yeon
    • The KIPS Transactions:PartC
    • /
    • v.9C no.6
    • /
    • pp.961-968
    • /
    • 2002
  • This paper proposes a new access method which is to considered the high variability in World Wide Web and manage the web cache space. Instead of using a single cache, we divide a cache and store all documents according to their file types. Proposed method was compares with current cache management policies using LFU, LRU and SIZE base algorithm. Using two different workload, we show the improvement hitting ratio and byte hitting ratio through simulating on the file type caching.

An Avoidance-Based Cache Consistency Algorithm without Unnecessary Callback (불필요한 콜백을 제거한 회피 기반의 캐쉬 일관성 알고리즘)

  • Kim, Chi-Yeon
    • Journal of Advanced Navigation Technology
    • /
    • v.10 no.2
    • /
    • pp.120-127
    • /
    • 2006
  • Data-caching of a client is an important technology for interaction with a server where data are cached and operated at the client node. Caching reduces network latency and increases resource utilization at the client. In a Client-server environment, the absence of cache consistency algorithm, we cannot guarantee the correctness of client applications. In this paper, we propose a new asynchronous avoidance-based cache consistency algorithm. We remove the additional callback due to the lock escalation in AACC. With a comprehensive performance analysis, we show that the proposed algorithm has less message exchange than the AACC. One-copy serializability is used for proving correctness of the proposed algorithm.

  • PDF

Web Caching Strategy based on Documents Popularity (선호도 기반 웹 캐싱 전략)

  • Yoo, Hae-Young;Park, Chel
    • Journal of KIISE:Computer Systems and Theory
    • /
    • v.29 no.9
    • /
    • pp.530-538
    • /
    • 2002
  • In this paper, we propose a new caching strategy for web servers. The proposed algorithm collects on]y the statistics of the requested file, for example the popularity, when a request arrives. And, at times, only files with higher popularity are cached all together. Because the cache remains unchanged until the cache is made newly, web server can use very efficient data structure for cache to determine whether a file is in the cache or not. This increases greatly tile efficiency of cache manipulation. Furthermore, the experiment that is performed with real log files built by web servers shows that the cache hit ratio and the cache hit ratio are better than those produced by LRU. The proposed algorithm has a drawback such that the cache hit ratio may decrease when the popularity of files that is not in the cache explodes instantaneously. But in our opinion, such explosion happens infrequently, and it is easy to implement the web servers to adapt them to such unusual cases.

Partition and Caching Mechanism for GML Visualization on Mobile Device (모바일 디바이스에서 GML 가시화를 위한 분할 및 캐싱 기법)

  • Song, Eun-Ha;Park, Yong-Jin;Han, Won-Hee;Jeong, Young-Sik
    • Journal of Korea Multimedia Society
    • /
    • v.11 no.7
    • /
    • pp.1025-1034
    • /
    • 2008
  • In this paper, we developed GridGML for efficiently supplying a GML and visualizing the map with partitioning map and caching method to a mobile device. In order to overcome the weighting of a file, which is the biggest weakness of a GML, GridGML extracts only the most necessary parts for the visualization of the map among GML attributes, and makes the file light as a class instance by applying an offset value. GridGML manages a partition based on the visualization area of a mobile device to visualize the map to a mobile device in real time, and transmits the partition area by serializing it for the benefit of transmission. Also, the received partition area is compounded in a mobile device and is visualized by being partitioned again as four visible areas based on the display of a mobile device. Then, the area is managed by applying a caching algorithm in consideration of repetitiveness for a received map for the efficient operation of resources. Also, in order to prevent the delay in transmission time as regards the instance density area of the map, an adaptive map partition mechanism is proposed for maintaining the transmission time uniformly.

  • PDF

Distributed Optimal Path Generation Based on Delayed Routing in Smart Camera Networks

  • Zhang, Yaying;Lu, Wangyan;Sun, Yuanhui
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.10 no.7
    • /
    • pp.3100-3116
    • /
    • 2016
  • With the rapid development of urban traffic system and fast increasing of vehicle numbers, the traditional centralized ways to generate the source-destination shortest path in terms of travel time(the optimal path) encounter several problems, such as high server pressure, low query efficiency, roads state without in-time updating. With the widespread use of smart cameras in the urban traffic and surveillance system, this paper maps the optimal path finding problem in the dynamic road network to the shortest routing problem in the smart camera networks. The proposed distributed optimal path generation algorithm employs the delay routing and caching mechanism. Real-time route update is also presented to adapt to the dynamic road network. The test result shows that this algorithm has advantages in both query time and query packet numbers.

Adaptive Mapping Information Management Scheme for High Performance Large Sale Flash Memory Storages (고성능 대용량 플래시 메모리 저장장치의 효과적인 매핑정보 캐싱을 위한 적응적 매핑정보 관리기법)

  • Lee, Yongju;Kim, Hyunwoo;Kim, Huijeong;Huh, Taeyeong;Jung, Sanghyuk;Song, Yong Ho
    • Journal of the Institute of Electronics and Information Engineers
    • /
    • v.50 no.3
    • /
    • pp.78-87
    • /
    • 2013
  • NAND flash memory has been widely used as a storage medium in mobile devices, PCs, and workstations due to its advantages such as low power consumption, high performance, and random accessability compared to a hard disk drive. However, NAND flash cannot support in-place update so that it is mandatory to erase the entire block before overwriting the corresponding page. In order to overcome this drawback, flash storages need a software support, named Flash Translation Layer. However, as the high performance mass NAND flash memory is getting widely used, the size of mapping tables is increasing more than the limited DRAM size. In this paper, we propose an adaptive mapping information caching algorithm based on page mapping to solve this DRAM space shortage problem. Our algorithm uses a mapping information caching scheme which minimize the flash memory access frequency based on the analysis of several workloads. The experimental results show that the proposed algorithm can increase the performance by up to 70% comparing with the previous mapping information caching algorithm.