• Title/Summary/Keyword: Cache consistency

Search Result 80, Processing Time 0.027 seconds

A Multiversion Timestamp Order Scheduling Method for Improving Commit Ratio of Mobile Transactions (이동 트랜잭션의 완료율 향상을 위한 다중버전 타임스탬프 순서화 스케줄링 기법)

  • Kim, Chi-Yeon;Hwang, Bu-Hyeon
    • The Transactions of the Korea Information Processing Society
    • /
    • v.6 no.5
    • /
    • pp.1143-1152
    • /
    • 1999
  • A Mobile computing environment is a new paradigm which provides users with the access to information irrespective of users' location. A transaction scheduling method for the mobile computing environment must be designed so that database consistency is maintained with considering the limited performance of mobile hosts and the narrow bandwidth of a wireless network. The scheduling method using a lock has some problems: the high message overhead between a server and a mobile host for maintaining a lock and the high abort ratio of the mobile transactions owing to violating the serializability when a mobile host uses a cache. So, in this paper, we propose an efficient transaction management method using timestamp to resolve these problems. The proposed method is used in the environment under which a mobile host uses a cache having two versions for each cached data item and a server maintains several versions for each data item. So, even though a mobile transaction is executed during several broadcasting interval, can be committed. As a result, the proposed method can improve the commit ratio of the mobile transactions by maintaining multiversion for each data item and does not require the additional message exchange to schedule transactions by using timestamp.

  • PDF

An Optimized Cache Coherence Protocol in Multiprocessor System Connected by Slotted Ring (슬롯링으로 연결된 다중처리기 시스템에서 최적화된 캐쉬일관성 프로토콜)

  • Min, Jun-Sik;Chang, Tae-Mu
    • The Transactions of the Korea Information Processing Society
    • /
    • v.7 no.12
    • /
    • pp.3964-3975
    • /
    • 2000
  • There are two policies for maintaining consistency among the multiple processor caches in a multiprocessor system: Write invalidate and Write update. In the write invalidate policy, whenever a processor attempt to write its cached block, it has to invalidate all the same copies of the updated block in the system. As a results of this frequent invalidations, this policy results in high cache miss ratio. On the other hand, the write update policy renew them, instead of invalidating all the same copies. This policy has to transfer the updated contents through interconnection network, whether the updated block is ptivate or not. Therefore the network suffer from heavy transaction traffic. In this paper we present an efficient cache coherence protocol for shared memory multiprocessor system connected by slotted ring. This protocol is based on the write update policy, but the updated contents are transferred only in case of updating the shared block. Otherwise, if the updated block is private, the updated contents are not transferred. We analyze the proposed protocol and enforce simulation to compare it with previous version.

  • PDF

Directory Cache Coherence Scheme using the Number-Balanced Binary Tree (수 평형 이진트리를 이용한 디렉토리 캐쉬 일관성 유지 기법)

  • Seo, Dae-Wha
    • The Transactions of the Korea Information Processing Society
    • /
    • v.4 no.3
    • /
    • pp.821-830
    • /
    • 1997
  • The directory-based cache coherence scheme is an attractive approach to solve the caceh coherence problem in a large-scale shared-memory multiprocessor.However, the exsting directory-based schemes have some problens such as the enormous storage overhead for a directory, the long invalidation latency, the heavy network condes-tion, and the low scalability.For resolving these problems, we propose a new directroy- based caceh coherence scheme which is suitable for building scalable, shred-memory multiprocessors.In this scheme, each directory en-try ofr a given memory block is a number-balanced binaty tree(NBBT) stucture.The NBBT has several proper-ties to effciently maintain the directory for the cache consistency such that the shape is unique, the maximum depth is [log$_2$n], and the tree has the minimum number of leaf nodes among the binarry tree with n nodes.Therefore, this scheme can reduce the storage overhead, the network traffic, and the inbalidation latency and can ensutr the high- scalability the large-scale shared-memory multiprocessors.

  • PDF

A Global Buffer Manager for a Shared Disk File System in SAN Clusters (SAN 환경에서 공유 디스크 파일 시스템을 위한 전역 버퍼 관리자)

  • 박선영;손덕주;신범주;김학영;김명준
    • Journal of KIISE:Computing Practices and Letters
    • /
    • v.10 no.2
    • /
    • pp.134-145
    • /
    • 2004
  • With rapid growth in the amount of data transferred on the Internet, traditional storage systems have reached the limits of their capacity and performance. SAN (Storage Area Network), which connects hosts to disk with the Fibre Channel switches, provides one of the powerful solutions to scale the data storage and servers. In this environment, the maintenance of data consistency among hosts is an important issue because multiple hosts share the files on disks attached to the SAN. To preserve data consistency, each host can execute the disk I/O whenever disk read and write operations are requested. However, frequent disk I/O requests cause the deterioration of the overall performance of a SAN cluster. In this paper, we introduce a SANtopia global buffer manager to improve the performance of a SAN cluster reducing the number of disk I/Os. We describe the design and algorithms of the SANtopia global buffer manager, which provides a buffer cache sharing mechanism among the hosts in the SAN cluster. Micro-benchmark results to measure the performance of block I/O operations show that the global buffer manager achieves speed-up by the factor of 1.8-12.8 compared with the existing method using disk I/O operations. Also, File system micro-benchmark results show that SANtopia file system with the global buffer manager improves performance by the factor of 1.06 in case of directories and 1.14 in case of files compared with the file system without a global buffer manager.

Cache Invalidation Schemes based on Time Guarantee for Improving Access Time in Mobile Ad hoc Networks (모바일 애드혹 네트워크에서 캐쉬 접근 시간 향상을 위한 시간보증 기반의 캐쉬무효화 기법)

  • Choi, Jae-Ho;Oh, Jae-Oh;Lee, Myong-Soo;Lee, Sang-Keun
    • The KIPS Transactions:PartC
    • /
    • v.16C no.1
    • /
    • pp.65-72
    • /
    • 2009
  • Due to the popularity of mobile devices and advances in wireless communication technologies, a mobile ad hoc network has received a lot of attention. In the existing data replication management research, the use of a replica has been shown to be an efficient technique for improving data accessibility. However, to use a replica in ad hoc networks, the data consistency between the original data and the replica should be guaranteed. In the traditional research, a mobile node should check an original data whether the data is updated or not. However, It may be costly or sometimes impossible to check the original data. In the case of the time constraint applications, the checking cost can cause more serious problem. In this paper, we propose the time-guarantee based cache invalidation schemes for time constraint applications and the threshold based compensation method to enhance the time-guarantee based scheme. The proposed schemes can remove the "rollback" problem. Simulation results show that our schemes outperform the previous ones in terms of access time with little loss of data currency.

A Cooperation with Clients on Agent-Initiated Cache Consistency Sheme (대리인에 의한 캐쉬 일관성 유지 방식에서의 클라이언트들간의 상호 협조)

  • 김경자;장태무
    • Proceedings of the Korean Information Science Society Conference
    • /
    • 1998.10a
    • /
    • pp.178-180
    • /
    • 1998
  • 분산 파일 시스템에서는 파일 서버는 사용률이 높고, 통신 네트워크나 클라이언트 워크스테이션은 상대적으로 낮은 것으로 보고 하고 있다. 따라서, 시스템의 확장은 파일 서버가 처리해야 할 작업 부담을 증가하는 결과를 초래하므로 파일 서버는 시스템 사의 주된 병목 구간이 된다. 따라서 분산 시스템의 성능을 향상시키기 위해서는 파일 서버의 작업 부담을 클리언트로 이전시키고 가급적이면 파일 서버가 유지되는 상태 정보의 양을 줄여야 한다. 본 논문에서는 파일 서버에서 유지되는 상태 정보의 양이 파일 서버의 작업 부담에 비례한다는 가정 하에 파일 서버의 작업부담을 클라이언트들에게 이양시키는 대리인에 의한 캐쉬 일관성 유지 방식과 클라이언트들간의 상호 협조를 통해 요구된 파일에 대한 응답시간을 줄이는 방안을 제안하였다.

  • PDF

Selective Cache Consistency Maintenance Schemes Considering the Connection States in Mobile Computing Environments (이동 컴퓨팅 환경에서 접속 상태를 고려한 선택적 캐쉬 일관성 유지 기법)

  • Kim, Hee-Sook;Hwang, Byung-Yeon
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2003.05b
    • /
    • pp.779-782
    • /
    • 2003
  • 이동 컴퓨팅 환경에서 이동 호스트가 사용하는 캐쉬는 이동 호스트의 자치성을 증대시킴으로서 전력의 한계와 제한된 대역폭이라는 단점을 보완할 수 있는 중요한 장치로 사용된다. 하지만 서버 데이터와의 일관성 유지를 해야하는 부담을 가지게 된다. 본 연구에서는 기존의 일관성 유지 기법이 가지는 불필요한 데이터의 방송 문제점을 보완하고 오랜 접속 단절 후에 연결된 이동 호스트에 대해서는 서버가 보관하고 있던 지난 방송 정보를 전달함으로서 대역폭을 효율적으로 이용하면서 이동 호스트의 자치성을 증대시키는 캐쉬 유지 기법을 제안한다.

  • PDF

A Modeling & Implementation of Client for Performance Evaluation of Cache Consistency Maintenance Algorithms (캐시 일관성 유지 알고리즘의 성능평가를 위한 Client 모델링 및 구현)

  • Shin, Seoung-Chul
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2007.05a
    • /
    • pp.53-56
    • /
    • 2007
  • CAD와 같이 계산위주의 응용을 지원하기 위해 개발된 초기의 객체지향 데이터베이스 시스템은 클라이언트에서 자료의 요구와 처리를 허용하는 자료 탑재방식이다. 이 방식은 클라이언트 워크스테이션에서 수행되는 작업의 대부분을 수행할 수 있는 지역 DBMS를 허용한다. 클라이언트 워크스테이션에서 자료와 잠금을 캐싱하는 것은 클라이언트-서버 데이터베이스 시스템의 성능 향상을 위한 중요한 기술이다. 이 논문에서는 클라이언트 워크스테이션이 자신의 지역 데이터베이스를 유지하면서 회피 또는 탐지 기반의 캐시일관성 유지 알고리즘들의 성능을 평가할 수 있도록 하는 클라이언트 워크스테이션을 모델링하고 구현하였다. 이 시스템은 3개의 클라이언트로 구성되고, 각 지역 동시성 제어 기법으로는 2PL을 사용하였다. 이 모델은 멀티데이터베이스 시스템 환경으로 확장 가능하다.

  • PDF

Memory Latency Hiding Techniques (메모리 지연을 감추는 기법들)

  • Ki, An-Do
    • Electronics and Telecommunications Trends
    • /
    • v.13 no.3 s.51
    • /
    • pp.61-70
    • /
    • 1998
  • The obvious way to make a computer system more powerful is to make the processor as fast as possible. Furthermore, adopting a large number of such fast processors would be the next step. This multiprocessor system could be useful only if it distributes workload uniformly and if its processors are fully utilized. To achieve a higher processor utilization, memory access latency must be reduced as much as possible and even more the remaining latency must be hidden. The actual latency can be reduced by using fast logic and the effective latency can be reduced by using cache. This article discusses what the memory latency problem is, how serious it is by presenting analytical and simulation results, and existing techniques for coping with it; such as write-buffer, relaxed consistency model, multi-threading, data locality optimization, data forwarding, and data prefetching.

Implementation of Integrated CPU-GPU for Efficient Uniform Memory Access Method and Verification System (CPU-GPU간 긴밀성을 위한 효율적인 공유메모리 접근 방법과 검증 시스템 구현)

  • Park, Hyun-moon;Kwon, Jinsan;Hwang, Tae-ho;Kim, Dong-Sun
    • IEMEK Journal of Embedded Systems and Applications
    • /
    • v.11 no.2
    • /
    • pp.57-65
    • /
    • 2016
  • In this paper, we propose a system for efficient use of shared memory between CPU and GPU. The system, called Fusion Architecture, assures consistency of the shared memory and minimizes cache misses that frequently occurs on Heterogeneous System Architecture or Unified Virtual Memory based systems. It also maximizes the performance for memory intensive jobs by efficient allocation of GPU cores. To test between architectures on various scenarios, we introduce the Fusion Architecture Analyzer, which compares OpenMP, OpenCL, CUDA, and the proposed architecture in terms of memory overhead and process time. As a result, Proposed fusion architectures show that the Fusion Architecture runs benchmarks 55% faster and reduces memory overheads by 220% in average.