• Title/Summary/Keyword: Update Transaction

Search Result 62, Processing Time 0.028 seconds

Transaction Management Using Update Protocol in Fully Replicated Databases (완전중복 데이터베이스에서 갱신 규약을 적용한 거래 관리)

  • Choe, Hui-Yeong;Hwang, Bu-Hyeon
    • The KIPS Transactions:PartD
    • /
    • v.9D no.1
    • /
    • pp.11-20
    • /
    • 2002
  • We propose a new concurrency control algorithm for enhancing the degree of parallelism of the transactions in fully replicated databases designed to improve the availability and the reliability. The update transactions can be logically decomposed of a read-only transaction and a write-only transaction. In our algorithm, a set of read operations of an update transaction is treated as a read-only transaction and the read-only transaction reads data items in the site to which it is submitted. And a set of write operations of the update transaction is treated as a write-only transaction and it is submitted to all corresponding sites after the update transaction has been completed. By using the proposed update protocol, all write-only transactions can execute at all sites atomically. The proposed algorithm can have transactions execute concurrently at the site to which they are submitted and, after the completion of each transaction, the update protocol is performed for updating their data items and checking their serializability. Therefore, the degree of parallelism of the transactions can be improved. Especially, if the probability of conflict among transactions is low, we can expect the higher degree of their parallelism.

Optimistic Concurrency Control with Update Transaction First for Broadcast Environment : OCC/UTF (방송환경에서 갱신 거래 우선 낙관적 동시성 제어 기법)

  • Lee, Uk-Hyeon;Hwang, Bu-Hyeon
    • The KIPS Transactions:PartD
    • /
    • v.9D no.2
    • /
    • pp.185-194
    • /
    • 2002
  • Most of mobile computing systems allow mostly read-only transactions from mobile clients for retrieving various types of Information such as stock data, traffic information and news updates. Since previous concurrence control protocols, however, do not consider such a particular characteristics, the performance degradation occurs when previous schemes are applied to the broadcast environment. In this paper, we propose OCC/UTF(Optimistic Concurrence Control with Update Transaction First) that is most appropriate for broadcast environment. OCC/UTF lets a query transaction, that has already read the data item which was invalidated by update transaction, read again the same data item without the abort of the query transaction due to non-serializability. Therefore, serializable order is maintained and the query transaction is committed safely regardless of commitment of update transactions. In OCC/UTF, Clients need not require server to commit their query transactions. Because of broadcasting the validation reports including values updated recently to clients, it reduces the overhead of requesting recent values from the server and the server need not also re-broadcast the newest values. As a result, OCC/UTF makes full use of the asymmetric bandwidth. It can also improve transaction throughput by increasing the commit ratio of query transactions as much as possible.

Database Construction for Design of the Components Software by Using an Incremental Update Propagation

  • Oh, Am-Suk;Kwon, Oh-Hyun
    • Journal of Korea Multimedia Society
    • /
    • v.6 no.4
    • /
    • pp.583-593
    • /
    • 2003
  • Engineering design applications require the support of long transactions in cooperative environments. The problem of the existing copy/update/merge approaches is that the partial effects of a committed transaction may be not part of the merged version. This paper introduces a new cooperative transaction model, which allows updates to be progressively notified or propagated into other transactions accessing the same object. To support incremental update propagation and notification, we use the term dynamic dependency to define the intertransaction dependency relationships among all the objects checked out from the public database. Consistency in multiple copies of the same object is achieved by a two-phase delta-merge protocol. Our model provides a synchronization of cooperative updates performed in several workspaces without using locking mechanisms.

  • PDF

Transaction Update method based on Fuzzy Chunk for QoS Performance Improvement of Mobile Streaming Service (모바일 스트리밍 서비스의 QoS 개선을 위한 퍼지 청크 기반의 트랜잭션 갱신방법)

  • Jeong, Taeg-Won;Lee, Chong-Deuk
    • Journal of Digital Contents Society
    • /
    • v.9 no.4
    • /
    • pp.543-550
    • /
    • 2008
  • There exist many QoS problems such as delay and jitter for mobile streaming due to limited bandwidth. This paper proposed a transaction update method based on Fuzzy chunk to solve the problems. The proposed method updated a transaction, which is classified as a read or an update, using Fuzzy filtering. A chunk block is rearranged according to the fuzziness which is classified as better relevance for fuzziness greater than 0.8, relevance for fuzziness between 0.5 and 0.7, and less irrelevance for fuzziness less than 0.5. According to the simulation result for the rearranged chunk block, the proposed method has better performance than those of RBM(Randon Based method), DBM(Distance Based Method), and FBM(Frequency Based Method).

  • PDF

A Design of Update Propagation Algorithm for Replica Consistency of Transaction (트랜잭션의 복제 일관성을 위한 갱신 전파 알고리즘 설계)

  • Lee Byung-Wook
    • Journal of Internet Computing and Services
    • /
    • v.4 no.6
    • /
    • pp.87-93
    • /
    • 2003
  • Data Replication makes distributed system and database system improved in availability and performance. But it is difficult to keep strict consistency in data update. Existing algorithms keep strict consistency, but have problems of cost a great deal and delay time. It is to introduce hybrid update propagation algorithm which permits eager update for original copy and propagates lazy update for other copies. Hybrid update propagation algorithm also permits group update in ownership. This algorithm menages replica version in order to control group update consistency. Consistency and performance is improved by combining eager update and lazy update with group ownership. Simulation shows improvement of transaction throughput and response time depending on application and execution environment.

  • PDF

Design and Implementation of a Reprocessing Transaction Model for Mobile Computing Environments (모바일 컴퓨팅 환경을 위한 재수행 트랜잭션 모델의 설계 및 구현)

  • 김동현;홍봉희
    • Journal of KIISE:Databases
    • /
    • v.30 no.2
    • /
    • pp.184-196
    • /
    • 2003
  • Mobile transactions updating spatial objects are long transactions that can update independently local objects of mobile clients during disconnection. Validation based protocols that are well known to be appropriate for mobile transactions are required to abort a conflicted transaction in order to keep data consistent. Since abortion leads to cancel of all of tile updates, it is not desirable to use the abortion for resolving conflicts of mobile transactions. In this paper, we propose a reprocessing transaction model to resolve the update conflicts between mobile transactions without aborting them. We also design a mobile transaction server to support the reprocessing transaction and build a prototype of a mobile field system. The reprocessing transaction is a subtransaction of a newly committed mobile transaction and re-executes only conflicted objects with foreign conflicted objects. We also introduce a progressive reprocessing scheme to expose non-conflicted objects of the mobile transaction to other transactions in order to reduce starvation of reprocessing transactions.

An Update Transaction Scheduling Method Using an Adaptive Request Message in Mobile Computing Environments (이동 컴퓨팅 환경에서 적응적 요청 메시지를 이용한 갱신 트랜잭션 스케줄링 기법)

  • Park, Jun;Chae, Duk-Jin;Hwang, Boo-Hyun;Kim, Joong-Bae;Jung, Seung-Wook
    • The KIPS Transactions:PartD
    • /
    • v.11D no.2
    • /
    • pp.339-350
    • /
    • 2004
  • A broadcast method in mobile computing environments which have very large client populations solves the problem of low bandwidth. But most of previous proposed concurrency control protocols are restricted to read-only transactions from mobile client. In this paper, we propose the UTSM-ARM method which uses caching and broadcast method, and allows update transactions in mobile client. The UTSM-ARM decides the validation of cache data consistency with the dynamic update pattern ratio of accessed cached data. Also, the timestamps of accessed cached data and transaction in mobile client guarantee the serializable execution of mobile transactions. As a result, UTSM-ARM makes efficiently use of the asymmetric bandwidth of broadcast environment and can Increase the transaction throughput.

Metadata Management for E-Commerce Transactions in Digital Library (디지털 도서관에서 전자상거래 트랜잭션을 위한 메타데이타 관리 기법)

  • Choe, Il-Hwan;Park, Seog
    • Journal of KIISE:Databases
    • /
    • v.29 no.1
    • /
    • pp.34-43
    • /
    • 2002
  • Since traditional static metadata set like Dublin Core has static metadata attributes about bibliography information, integration of metadata for various metadata, problems about standard and extension of metadata must be considered for applying it to new environment. Specially, as event-driven metadata write method included the notion of e-commerce come out for interoperability in digital libraries, traditional metadata management which cannot distinguish between different kinds of update operations to new extension of metadata set occurs unsuitable waiting of update operation. So, improvement is needed about it. In this paper, we show whether alleviative transaction consistency can be applied to digital library or not. Also it would divide newer metadata into static metadata attribute connected in read operation within user read-only transaction and dynamic metadata attribute in update operation within dynamic(e-commerce) update transactions. We propose newer metadata management algorithm considered in classfication of metadata attributes and dynamic update transaction. Using two version for minimal maintenance cost and ARU(Appended Refresh Unit) for dynamic update transaction, to minimize conflict between read and write operations shows fast response time and high recency ratio. As a result of the performance evaluation, we show our algorithm is proved to be better than other algorithms in newer metadata environments.

A Study for Efficient Transaction Management in Broadcast Environments (방송 환경에서 효율적인 트랜잭션 관리 방법에 관한 연구)

  • 김치연
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2002.11a
    • /
    • pp.648-651
    • /
    • 2002
  • In data broadcast environments, data is disseminated from a small number of sewers to a much larger number of clients. The advancing technology of hardware and increasing needs of users' have extended a broadcast environments. A mobile environment use the data broadcasting. A user carrying a portable computer can submit the operations of a transaction. Some limitations such as the poor-resource, user mobility, and low bandwidth are difficult to apply traditional transaction management to the mobile environments. So, the new mechanisms are needed and one of the them is the correctness criteria of transactions. Serializability is the most used criteria, but serializability is not appropriate to the broadcast environment because unnecessary abort and additional message exchanges are occurred. Hence, in this paper, we will address the need of the new correctness criteria weaker than serializability and describe the adoption of the update consistency.

  • PDF

An Optimal Denormalization Method in Relational Database with Response Time Constraint (관계형 데이터베이스에서 응답시간에 제약이 있는 경우 최적 역정규화 방법)

  • 장영관
    • Journal of Korean Society of Industrial and Systems Engineering
    • /
    • v.26 no.3
    • /
    • pp.1-9
    • /
    • 2003
  • Databases are central to business information systems and RDBMS is most widely used for the database system. Normalization was designed to control various anomalies (insert, update, delete anomalies). However, normalized database design does not account for the tradeoffs necessary for performance. In this research, I model a database design method considering denormalization of duplicating attributes in order to reduce frequent join processes. This model considers response time for processing each select, insert, update, delete transaction, and for treating anomalies. A branch and bound method is proposed for this model.