• Title/Summary/Keyword: data caching

Search Result 222, Processing Time 0.024 seconds

Performance Evaluation of Deferrd Locking for Maintaining Transactional Cache Consistency (트랜잭션 캐쉬 일관성을 유지하기 위한 지연 로킹 기법의 성능 평가)

  • Kwon, Hyeok-Min
    • The Transactions of the Korea Information Processing Society
    • /
    • v.7 no.8
    • /
    • pp.2310-2326
    • /
    • 2000
  • Client-server DBMS based on a data-shipping model can exploit e1ient resources effectively by allowing inter-transaction caching. However, inter-transaction caching raises the need of transactional cache consistency maintenancetTCCM protocol. since each client is able to cache a portion of the database dynamically. Deferred locking(DL) is a new detection-based TCCM scheme designed on the basis of a primary copy locking algorithm. In DL, a number of lock ,ujuests and a data shipping request are combined into a single message packet to minimize the communication overhead required for consistency checking. Lsing a simulation model. the performance of the prolxlsed scheme is compared with those of two representative detection based schemes, the adaptive optimistic concurrency control and the caching two-phase locking. The performance results indicate that DL improves the overall system throughput with a reasonable transaction abort ratio over other detection - based schemes.

  • PDF

Semantic schema data processing using cache mechanism (캐쉬메카니즘을 이용한 시맨틱 스키마 데이터 처리)

  • Kim, Byung-Gon;Oh, Sung-Kyun
    • Journal of the Korea Society of Computer and Information
    • /
    • v.16 no.3
    • /
    • pp.89-97
    • /
    • 2011
  • In semantic web information system like ontology that access distributed information from network, efficient query processing requires an advanced caching mechanism to reduce the query response time. P2P network system have become an important infra structure in web environment. In P2P network system, when the query is initiated, reducing the demand of data transformation to source peer is important aspect of efficient query processing. Caching of query and query result takes a particular advantage by adding or removing a query term. Many of the answers may already be cached and can be delivered to the user right away. In web environment, semantic caching method has been proposed which manages the cache as a collection of semantic regions. In this paper, we propose the semantic caching technique in cluster environment of peers. Especially, using schema data filtering technique and schema similarity cache replacement method, we enhanced the query processing efficiency.

Femto-Caching File Placement Technique for Overlapped Helper Coverage Without User Location Information (사용자 위치정보를 사용하지 않는 헬퍼 간 중첩 커버리지 영역을 위한 펨토-캐싱 파일 분배 기술)

  • Shim, Jae-Nam;Min, Byoung-Yoon;Kim, Dong Ku
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.39A no.11
    • /
    • pp.682-689
    • /
    • 2014
  • Due to explosive growth of mobile data traffic, many kind of techniques based on small cell is proposed as solution for phenomenon. However, those techniques essentially demands increase of backhaul capacity and causes performance degradation if not satisfied. Based on that, the approach applying the storage capacity in place of backhaul capacity, which is known as femto-caching, is proposed to reduce data downloading delay of users in system. In this paper, we expanded previous research by proposing file placement strategy with distribution of user position, which is more practical scenario. Simulation results verify that our proposed scheme has better performance gains mainly because when coverage of helpers are overlapped, users get more opportunity to connect various helpers which enables users to download a variety kind of files from helpers, not base station.

Location-Based Services for Dynamic Range Queries

  • Park Kwangjin;Song Moonbae;Hwang Chong-Sun
    • Journal of Communications and Networks
    • /
    • v.7 no.4
    • /
    • pp.478-488
    • /
    • 2005
  • To conserve the usage of energy, indexing techniques have been developed in a wireless mobile environment. However, the use of interleaved index segments in a broadcast cycle increases the average access latency for the clients. In this paper, we present the broadcast-based location dependent data delivery scheme (BBS) for dynamic range queries. In the BBS, broadcasted data objects are sorted sequentially based on their locations, and the server broadcasts the location dependent data along with an index segment. Then, we present a data prefetching and caching scheme, designed to reduce the query response time. The performance of this scheme is investigated in relation to various environmental variables, such as the distributions of the data objects, the average speed of the clients, and the size of the service area.

A Study on Data Availability Improvement using Mobility Prediction Technique with Location Information (위치 정보와 이동 예측 기법을 이용한 데이터 가용성 향상에 관한 연구)

  • Yang, Hwan Seok
    • Journal of Korea Society of Digital Industry and Information Management
    • /
    • v.8 no.4
    • /
    • pp.143-149
    • /
    • 2012
  • MANET is a network that is a very useful application to build network environment in difficult situation to build network infrastructure. But, nodes that configures MANET have difficulties in data retrieval owing to resources which aren't enough and mobility. Therefore, caching scheme is required to improve accessibility and availability for frequently accessed data. In this paper, we proposed a technique that utilize mobility prediction of nodes to retrieve quickly desired information and improve data availability. Mobility prediction of modes is performed through distance calculation using location information. We used technique which global cluster table and local member table is managed by cluster head to reduce data consistency and query latency time. We compared COCA and CacheData and experimented to confirm performance of proposed scheme in this paper and efficiency of the proposed technique through experience was confirmed.

Distributed In-Memory Caching Method for ML Workload in Kubernetes (쿠버네티스에서 ML 워크로드를 위한 분산 인-메모리 캐싱 방법)

  • Dong-Hyeon Youn;Seokil Song
    • Journal of Platform Technology
    • /
    • v.11 no.4
    • /
    • pp.71-79
    • /
    • 2023
  • In this paper, we analyze the characteristics of machine learning workloads and, based on them, propose a distributed in-memory caching technique to improve the performance of machine learning workloads. The core of machine learning workload is model training, and model training is a computationally intensive task. Performing machine learning workloads in a Kubernetes-based cloud environment in which the computing framework and storage are separated can effectively allocate resources, but delays can occur because IO must be performed through network communication. In this paper, we propose a distributed in-memory caching technique to improve the performance of machine learning workloads performed in such an environment. In particular, we propose a new method of precaching data required for machine learning workloads into the distributed in-memory cache by considering Kubflow pipelines, a Kubernetes-based machine learning pipeline management tool.

  • PDF

Client-Side Caching for Nearest Neighbor Queries

  • Park Kwangjin;Hwang Chong-Sun
    • Journal of Communications and Networks
    • /
    • v.7 no.4
    • /
    • pp.417-428
    • /
    • 2005
  • The Voronoi diagram (VD) is the most suitable mechanism to find the nearest neighbor (NN) for mobile clients. In NN query processing, it is important to reduce the query response time, since a late query response may contain out-of-date information. In this paper, we study the issue of location dependent information services (LDISs) using a VD. To begin our study, we first introduce a broadcast-based spatial query processing methods designed to support NN query processing. In further sections, we introduce a generic method for location-dependent sequential prefetching and caching. The performance of this scheme is studied in different simulated environments. The core contribution of this research resides in our analytical proof and experimental results.

A Comparative Analysis on Page Caching Strategies Affecting Energy Consumption in the NAND Flash Translation Layer (NAND 플래시 변환 계층에서 전력 소모에 영향을 미치는 페이지 캐싱 전략의 비교·분석)

  • Lee, Hyung-Bong;Chung, Tae-Yun
    • IEMEK Journal of Embedded Systems and Applications
    • /
    • v.13 no.3
    • /
    • pp.109-116
    • /
    • 2018
  • SSDs that are not allowed in-place update within the allocated page cause another allocation of a new page that will replace the previous page at the moment data modification occurs. This intrinsic characteristic of SSDs requires many changes to the existing HDD-based IO theory. In this paper, we conduct a performance comparison of FTL caching strategy in perspective of cache hashing (Global vs. grouped) and caching algorithm (LRU vs. NUR) through a simulation. Experimental results show that in terms of energy consumption for flash operation the grouped management of cache is not suitable and NUR algorithm is superior to LRU algorithm. In particular, we found that the cache hit ratio of LRU algorithm is about 10% point higher than that of NUR algorithm while the energy consumption of LRU algorithm is about 32% high.

Proxy Caching Mechanism for Continuous Media Data on the Internet (인터넷 상의 연속 미디어를 위한 Proxy Caching 기법)

  • 임은지;최태욱;박성호;정기동
    • Proceedings of the Korean Information Science Society Conference
    • /
    • 2000.04a
    • /
    • pp.382-384
    • /
    • 2000
  • WWW의 성장으로 인터넷은 과부하 상태에 이르렀으며, 현재 인터넷 상에는 오디오나 비디오 같은 연속미디어 데이터가 급격히 증가하는 추세에 있다. 본 논문은 인터넷상의 연속미디어 데이터의 효율적인 서비스를 위한 프락시 캐슁 기법을 제안한다. 연속미디어 스트림의 앞부분부터 점진적으로 캐슁하고, 스트림의 캐슁 길이를 유동적으로 변화시켜서 인기도에 따른 사용공간의 차별화를 수행한다. 그리하여 캐쉬의 제한된 저장공간을 보다 효율적으로 사용하고, 클라이언트의 서비스 지연시간을 최소화 시킬 수 있다. 그리고, 실험을 통하여 다른 알고리즘과의 성능을 비교 측정한다.

  • PDF

Game-Based Content Caching and Data Sponsor Scheme for the Content Network (콘텐츠 네트워크 환경에서 게임이론을 이용한 콘텐츠 캐싱 및 데이터 스폰서 기법)

  • Won, JoongSeop;Kim, SungWook
    • KIPS Transactions on Computer and Communication Systems
    • /
    • v.8 no.7
    • /
    • pp.167-176
    • /
    • 2019
  • Recently, as the types of services that can be enjoyed in mobile telecommunication networks such as social networks and video streaming are increasing, mobile users(MUs) can access mobile contents easily by consuming mobile data. However, under a mobile telecommunication environment, MUs have to pay a high data fee to a network service provider(SP) in order to enjoy contents. The 'data sponsor' technique, introduced as a way to solve this problem, has attracted attention as a breakthrough method for enhancing contents accessibility of MUs. In this paper, we propose an algorithm that determines the optimal discount rate through the Stackelberg game in the data sponsor environment. We also propose an algorithm to design edge caching, which caches highly popular content for MUs on edge server, through many-to-many matching game. Simulation results clearly indicate that the profit for CP's content consumption is improved by about 6~11%, and the profit of CP according to the ratio of edge caching is improved by about 12% than the other existing schemes under data sponsor environment.