• Title/Summary/Keyword: 캐싱

Search Result 280, Processing Time 0.032 seconds

Cache Management Method for Query Forwarding Optimization in the Grid Database (그리드 데이터베이스에서 질의 전달 최적화를 위한 캐쉬 관리 기법)

  • Shin, Soong-Sun;Jang, Yong-Il;Lee, Soon-Jo;Bae, Hae-Young
    • Journal of Korea Multimedia Society
    • /
    • v.10 no.1
    • /
    • pp.13-25
    • /
    • 2007
  • A cache is used for optimization of query forwarding in the Grid database. To decrease network transmission cost, frequently used data is cached from meta database. Existing cache management method has a unbalanced resource problem, because it doesn't manage replicated data in each node. Also, it increases network cost by cache misses. In the case of data modification, if cache is not updated, queries can be transferred to wrong nodes and it can be occurred others nodes which have same cache. Therefore, it is necessary to solve the problems of existing method that are using unbalanced resource of replica and increasing network cost by cache misses. In this paper, cache management method for query forwarding optimization is proposed. The proposed method manages caches through cache manager. To optimize query forwarding, the cache manager makes caching data from lower loaded replicated node. The query processing cost and the network cost will decrease for the reducing of wrong query forwarding. The performance evaluation shows that proposed method performs better than the existing method.

  • PDF

File-System-Level SSD Caching for Improving Application Launch Time (응용프로그램의 기동시간 단축을 위한 파일 시스템 수준의 SSD 캐싱 기법)

  • Han, Changhee;Ryu, Junhee;Lee, Dongeun;Kang, Kyungtae;Shin, Heonshik
    • Journal of KIISE
    • /
    • v.42 no.6
    • /
    • pp.691-698
    • /
    • 2015
  • Application launch time is an important performance metric to user experience in desktop and laptop environment, which mostly depends on the performance of secondary storage. Application launch times can be reduced by utilizing solid-state drive (SSD) instead of hard disk drive (HDD). However, considering a cost-performance trade-off, utilizing SSDs as caches for slow HDDs is a practicable alternative in reducing the application launch times. We propose a new SSD caching scheme which migrates data blocks from HDDs to SSDs. Our scheme operates entirely in the file system level and does not require an extra layer for mapping SSD-cached data that is essential in most other schemes. In particular, our scheme does not incur mapping overheads that cause significant burdens on the main memory, CPU, and SSD space for mapping table. Experimental results conducted with 8 popular applications demonstrate our scheme yields 56% of performance gain in application launch, when data blocks along with metadata are migrated.

User Centric Cache Allocation Schemes in Infrastructure Wireless Mesh Networks (인프라스트럭처 무선 메쉬 네트워크에서 사용자 중심 캐싱 할당 기법)

  • Jeon, Seung Hyun
    • Journal of Industrial Convergence
    • /
    • v.17 no.4
    • /
    • pp.131-137
    • /
    • 2019
  • In infrastructure wireless mesh networks (WMNs), in order to improve mobile users' satisfaction for the given cache hit ratio, we investigate an User centric Cache Allocation (UCA) scheme while reducing cache cost in a mesh router (MR) and expected transmission time (ETT) for content search in cache. To minimize ETT values of mobile users, a genetic algorithm based UCA (GA-UCA) scheme is provided. The goal is to maximize mobile users' satisfaction via our well defined utility, which considers content popularity and the number of mobile users. Finally, through solving optimization problem we show the optimal cache can be allocated for UCA and GA-UCA. Besides, a WMN provider can find the optimal number of mobile users for user centric cache allocation in infrastructure WMNs.

A Mapping Table Caching Scheme for NAND Flash-based Mobile Storage Devices (NAND 플래시 기반 모바일 저장장치를 위한 사상 테이블 캐싱 기법)

  • Yang, Soo-Hyeon;Ryu, Yeon-Seung
    • The Journal of Society for e-Business Studies
    • /
    • v.15 no.4
    • /
    • pp.21-31
    • /
    • 2010
  • Recently e-business such as online financial trade and online shopping using mobile computes are widely spread. Most of mobile computers use NAND flash memory-based storage devices for storing data. Flash memory storage devices use a software called flash translation layer to translate logical address from a file system to physical address of flash memory by using mapping tables. The legacy FTLs have a problem that they must maintain very large mapping tables in the RAM. In order to address this issues, in this paper, we proposed a new caching scheme of mapping tables. We showed through the trace-driven simulations that the proposed caching scheme reduces the space overhead dramatically but does not increase the time overhead. In the case of online transaction workload in e-business environment, in particular, the proposed scheme manifests better performance in reducing the space overhead.

Caching and Prefetching Policies Using Program Page Reference Patterns on a File System Layer for NAND Flash Memory (NAND 플래시 메모리용 파일 시스템 계층에서 프로그램의 페이지 참조 패턴을 고려한 캐싱 및 선반입 정책)

  • Park, Sang-Oh;Kim, Kyung-San;Kim, Sung-Jo
    • The KIPS Transactions:PartA
    • /
    • v.14A no.4
    • /
    • pp.235-244
    • /
    • 2007
  • Caching and prefetching policies have been used in most of computer systems to compensate speed differences between primary memory and secondary storage devices. In this paper, we design and implement a Flash Cache Core Module(FCCM) on the YAFFS which operates on a file system layer for NAND flash memory. The FCCM is independent of the underlying kernel in order to support its stability and compatibility. Also, we implement the Dirty-Last memory replacement technique considering the characteristics of flash memory, and the waiting queue for pages to be prefetched according to page hit. The FCCM reduced the number of I/Os and the amount of prefetched pages by maximum 55%(20% on average) and maximum 55%(24% on average), respectively, comparing with caching and prefetching policies of Linux.

Partition and Caching Mechanism for GML Visualization on Mobile Device (모바일 디바이스에서 GML 가시화를 위한 분할 및 캐싱 기법)

  • Song, Eun-Ha;Park, Yong-Jin;Han, Won-Hee;Jeong, Young-Sik
    • Journal of Korea Multimedia Society
    • /
    • v.11 no.7
    • /
    • pp.1025-1034
    • /
    • 2008
  • In this paper, we developed GridGML for efficiently supplying a GML and visualizing the map with partitioning map and caching method to a mobile device. In order to overcome the weighting of a file, which is the biggest weakness of a GML, GridGML extracts only the most necessary parts for the visualization of the map among GML attributes, and makes the file light as a class instance by applying an offset value. GridGML manages a partition based on the visualization area of a mobile device to visualize the map to a mobile device in real time, and transmits the partition area by serializing it for the benefit of transmission. Also, the received partition area is compounded in a mobile device and is visualized by being partitioned again as four visible areas based on the display of a mobile device. Then, the area is managed by applying a caching algorithm in consideration of repetitiveness for a received map for the efficient operation of resources. Also, in order to prevent the delay in transmission time as regards the instance density area of the map, an adaptive map partition mechanism is proposed for maintaining the transmission time uniformly.

  • PDF

Game-Based Content Caching and Data Sponsor Scheme for the Content Network (콘텐츠 네트워크 환경에서 게임이론을 이용한 콘텐츠 캐싱 및 데이터 스폰서 기법)

  • Won, JoongSeop;Kim, SungWook
    • KIPS Transactions on Computer and Communication Systems
    • /
    • v.8 no.7
    • /
    • pp.167-176
    • /
    • 2019
  • Recently, as the types of services that can be enjoyed in mobile telecommunication networks such as social networks and video streaming are increasing, mobile users(MUs) can access mobile contents easily by consuming mobile data. However, under a mobile telecommunication environment, MUs have to pay a high data fee to a network service provider(SP) in order to enjoy contents. The 'data sponsor' technique, introduced as a way to solve this problem, has attracted attention as a breakthrough method for enhancing contents accessibility of MUs. In this paper, we propose an algorithm that determines the optimal discount rate through the Stackelberg game in the data sponsor environment. We also propose an algorithm to design edge caching, which caches highly popular content for MUs on edge server, through many-to-many matching game. Simulation results clearly indicate that the profit for CP's content consumption is improved by about 6~11%, and the profit of CP according to the ratio of edge caching is improved by about 12% than the other existing schemes under data sponsor environment.

SSD Cache for RAID: Integrating Data Caching and Parity Update Delay (RAID를 위한 SSD 캐시: 데이터 캐싱과 패리티 갱신 지연 기법의 결합)

  • Minh, Sophal;Lee, Donghee
    • KIISE Transactions on Computing Practices
    • /
    • v.23 no.6
    • /
    • pp.379-385
    • /
    • 2017
  • In enterprise environments, hybrid storage typically utilizes SSDs over disk-based RAID. Typically, SSDs over RAID are used as the data cache. Recently, the LeavO caching scheme was introduced to reduce the parity update overhead of the underlying RAID. In this paper, we combine the data caching and LeavO caching schemes and derive cost models of the combined cache to determine the optimal data and LeavO cache sizes. We also propose the Adaptive Combined Cache that dynamically adjusts the data cache and LeavO cache sizes for evolving workloads. Experimental results show that the performance of the Adaptive Combined Cache is significantly superior to that of the conventional data caching scheme and is comparable with that of the off-line optimal scheme.

A Degraded Quality Service Policy for reducing the transcoding loads in a Transcoding Proxy (트랜스코딩 프록시에서 트랜스코딩 부하를 줄이기 위한 낮은 품질 서비스 정책)

  • Park, Yoo-Hyun
    • The KIPS Transactions:PartA
    • /
    • v.16A no.3
    • /
    • pp.181-188
    • /
    • 2009
  • Transcoding is one of core techniques that implement VoD services according to QoS. But it consumes a lot of CPU resource. A transcoding proxy transcodes multimedia objects to meet requirements of various mobile devices and caches them to reuse later. In this paper, we propose a service policy that reduces the load of transcoding multimedia objects by degrading QoS in a transcoding proxy. Due to the tradeoff between QoS and the load of a proxy system, a transcoding proxy provides lower QoS than a client's requirement so that it can accomodate more clients.

A Study on Caching Management Technique in Mobile Ad-hoc Network (Mobile Ad-hoc Network에서 캐싱 관리 기법에 관한 연구)

  • Yang, Hwan Seok;Yoo, Seung Jae
    • Convergence Security Journal
    • /
    • v.12 no.4
    • /
    • pp.91-96
    • /
    • 2012
  • MANET is developed technique fairly among many field of wireless network. Nodes which consist of MANET transmit data using multi-hop wireless connection. Caching scheme is technique which can improve data access capacity and availability of nodes. Previous studies were achieved about dynamic routing protocol to improve multi-hop connection of moving nodes. But management and maintenance of effective cache information because of movement of nodes is not easy. In this study, we proposed cluster-based caching scheme to manage connection by decreasing overhead and moving of nodes as moving node discovers cache of wish information. And HLP was used to maintain effective cache table in each cluster head. Efficiency of proposed technique in this study was confirmed by experiment.