• Title/Summary/Keyword: Transaction Log

Search Result 62, Processing Time 0.025 seconds

A Recovery Technique Using Client-based Logging in Client/Server Environment

  • Park, Yong-Mun;Lee, Chan-Seob;Kim, Dong-Hyuk;Park, Eui-In
    • Proceedings of the IEEK Conference
    • /
    • 2002.07a
    • /
    • pp.429-432
    • /
    • 2002
  • The existing recovery technique using the logging technique in the client/sewer database system only administers the log as a whole in a server. This contains the logging record transmission cost on the transaction that is executed in each client potentially and increases network traffic. In this paper, the logging technique for redo-only log is suggested, which removes the redundant before-image and supports the client-based logging to eliminate the transmission cost of the logging record. Also, in case of a client crash, redo recovery through a backward client analysis log is performed in a self-recovering way. In case of a server crash, the after-image of the pages which needs recovery through simultaneous backward analysis log is only transmitted and redo recovery is done through the received after-image and backward analysis log. Also, we select the comparing model to estimate the performance about the proposed recovery technique. And we analyzed the redo and recovery time about the change of the number of client and the rate of updating operation.

  • PDF

Disjunctive Process Patterns Refinement and Probability Extraction from Workflow Logs

  • Kim, Kyoungsook;Ham, Seonghun;Ahn, Hyun;Kim, Kwanghoon Pio
    • Journal of Internet Computing and Services
    • /
    • v.20 no.3
    • /
    • pp.85-92
    • /
    • 2019
  • In this paper, we extract the quantitative relation data of activities from the workflow event log file recorded in the XES standard format and connect them to rediscover the workflow process model. Extract the workflow process patterns and proportions with the rediscovered model. There are four types of control-flow elements that should be used to extract workflow process patterns and portions with log files: linear (sequential) routing, disjunctive (selective) routing, conjunctive (parallel) routing, and iterative routing patterns. In this paper, we focus on four of the factors, disjunctive routing, and conjunctive path. A framework implemented by the authors' research group extracts and arranges the activity data from the log and converts the iteration of duplicate relationships into a quantitative value. Also, for accurate analysis, a parallel process is recorded in the log file based on execution time, and algorithms for finding and eliminating information distortion are designed and implemented. With these refined data, we rediscover the workflow process model following the relationship between the activities. This series of experiments are conducted using the Large Bank Transaction Process Model provided by 4TU and visualizes the experiment process and results.

A Recovery Scheme of Single Node Failure using Version Caching in Database Sharing Systems (데이타베이스 공유 시스템에서 버전 캐싱을 이용한 단일 노드 고장 회복 기법)

  • 조행래;정용석;이상호
    • Journal of KIISE:Databases
    • /
    • v.31 no.4
    • /
    • pp.409-421
    • /
    • 2004
  • A database sharing system (DSS) couples a number of computing nodes for high performance transaction processing, and each node in DSS shares database at the disk level. In case of node failures in DSS, database recovery algorithms are required to recover the database in a consistent state. A database recovery process in DSS takes rather longer time compared with single database systems, since it should include merging of discrete log records in several nodes and perform REDO tasks using the merged lo9 records. In this paper, we propose a two version caching (2VC) algorithm that improves the cache fusion algorithm introduced in Oracle 9i Real Application Cluster (ORAC). The 2VC algorithm can achieve faster database recovery by eliminating the use of merged log records in case of single node failure. Furthermore, it can improve the performance of normal transaction processing by reducing the amount of unnecessary disk force overhead that occurs in ORAC.

Design and Implementation of the Surrogate Transaction Manager for Mobile GIS (모바일 GIS를 위한 대리 트랜잭션 관리자의 설계 및 구현)

  • 반재훈;문선희;김동현;홍봉희
    • Journal of KIISE:Databases
    • /
    • v.30 no.4
    • /
    • pp.397-407
    • /
    • 2003
  • Transactions of updating spatial dta with mobile clients are log transactions because a user disconnected from a server surveys real features and updates them. In this environment, it is appropriate to exploit the optimistic approach based on the validation test in order to control the concurrency of transactions. On the contrary, the pessimistic concurrency control scheme makes transactions wait for a long time due to the lock. In this paper, we propose the surrogate transaction model and implement its manager for the S-S-M(Server-Surrogate PC-Mobile Client) structure which is appropriate for updating spatial data in mobile environments. In the S-S-M structure, the mobile client communicates with the server by the surrogate PC. We extend the validation condition in consideration of spatial relationships between spatial objects in this model. We also present the commit protocol where the user of a surrogate PC adjusts objects of the conflicted surrogate transaction to minimize costs for the abortion of the transaction.

Using Transaction Logs to Better Understand User Search Session Patterns in an Image-based Digital Library (이미지 기반 디지털 도서관에서 이용자 검색 패턴의 효과적 이해를 위한 트랜잭션 로그 데이터 분석)

  • Han, Hye-Jung;Joo, Soohyung;Wolfram, Dietmar
    • Journal of the Korean BIBLIA Society for library and Information Science
    • /
    • v.25 no.1
    • /
    • pp.19-37
    • /
    • 2014
  • Server transaction logs containing complete click-through data from a digital library of primarily image-based documents were analyzed to better understand user search session behavior. One month of data was analyzed using descriptive statistics and network analysis methods. The findings reveal iterative search behaviors centered on result views and evaluation and topical areas of focus for the search sessions. The study is novel in its combined analytical techniques and use of click-through data for image collections.

Consistency preservation techniques for Location Register System in Mobile Networks

  • Kim, Jang-Hwan
    • International Journal of Internet, Broadcasting and Communication
    • /
    • v.12 no.2
    • /
    • pp.144-149
    • /
    • 2020
  • A database called Home Location Register(HLR) plays a major role in location management in mobile cellular networks. The objectives of this paper are to identify the problems of the current HLR system through rigorous analysis, to suggest solutions to them. The current HLR backup method is a process of simply writing the changed memory SLD block to disk, which has a problem in maintaining database consistency. Since information change and backup are performed separately by separate processes, there is a risk of information inconsistency when an error restart occurs. To solve this problem, a transaction concept was introduced for subscriber-related operation functions and a recovery method through logging and checkpointing was introduced. The subscriber related functions of tasks terminated normally by the suggested process are recovered with consistency even after system restarts. Performance is also not affected seriously because disk tasks for log occur with only subscriber related functions.

An Efficient Recovery Method for Mobile Main Memory Database System (모바일 메인메모리 데이터베이스 시스템을 위한 효율적인 복구 기법)

  • Cho, Sung-Je
    • Journal of Information Technology Services
    • /
    • v.7 no.2
    • /
    • pp.181-195
    • /
    • 2008
  • The rapid growth of mobile communication technology has provided the expansion of mobile internet services, particularly mobile realtime transaction takes much weight among mobile fields. There is an increasing demand for various mobile applications to process transactions in a mobile computing fields. Thus, During transmission in wireless networks a base station failure inevitably causes data loss of the base station buffer. It is required to compensate the loss for communication. The existing methods for a base station failure are not adequate because they all suffer from too much overhead and resolve only the link failure. In this paper, we study an efficient recovry systems for a mobile DBMS. We propose SLL (Segment Log List) that enables the mobile host to compensate data loss efficiently in the case of base station failure. In SLL, a base station deliveries an output information of data cells to a mobile host. when a base station fails, the mobile host can retransmit just next data cells. We also prove the efficiency of new method.

Real time Storage Manager to store very large datausing block transaction (블록 단위 트랜잭션을 이용한 대용량 데이터의 실시간 저장관리기)

  • Baek, Sung-Ha;Lee, Dong-Wook;Eo, Sang-Hun;Chung, Warn-Ill;Kim, Gyoung-Bae;Oh, Young-Hwan;Bae, Hae-Young
    • Journal of Korea Spatial Information System Society
    • /
    • v.10 no.2
    • /
    • pp.1-12
    • /
    • 2008
  • Automatic semiconductor manufacture system generating transaction from 50,000 to 500,000 per a second needs storage management system processing very large data at once. A lot of storage management systems are researched for storing very large data. Existing storage management system is typical DBMS on a disk. It is difficult that the DBMS on a disk processes the 500,000 number of insert transaction per a second. So, the DBMS on main memory appeared to use memory. But it is difficultthat very large data stores into the DBMS on a memory because of limited amount of memory. In this paper we propose storage management system using insert transaction of a block unit that can process insert transaction over 50,000 and store data on low storage cost. A transaction of a block unit can decrease cost for a log and index per each tuple as transforming a transaction of a tuple unit to a block unit. Besides, the proposed system come cost to decompress all block of data because the information of each field be loss. To solve the problems, the proposed system generates the index of each compressed block to prevent reducing speed for searching. The proposed system can store very large data generated in semiconductor system and reduce storage cost.

  • PDF

Fuzzy category based transaction analysis for web usage mining (웹 사용 마이닝을 위한 퍼지 카테고리 기반의 트랜잭션 분석 기법)

  • 이시헌;이지형
    • Proceedings of the Korean Institute of Intelligent Systems Conference
    • /
    • 2004.04a
    • /
    • pp.341-344
    • /
    • 2004
  • 웹 사용 마이닝(Web usage mining)은 웹 로그 파일(web log file)이나 웹 사용 데이터(Web usage data)에서 의미 있는 정보를 찾아내는 연구 분야이다. 웹 사용 마이닝에서 일반적으로 많이 사용하는 웹 로그 파일은 사용자들이 참조한 페이지의 단순한 리스트들이다. 따라서 단순히 웹 로그 파일만을 이용하는 방법만으로는 사용자가 참조했던 페이지의 내용을 반영하여 분석하는데에는 한계가 있다. 이러한 점을 개선하고자 본 논문에서는 페이지 위주가 아닌 웹 페이지가 포함하고 있는 내용(아이템)을 고려하는 새로운 퍼지 카테고리 기반의 웹 사용 마이닝 기법을 제시한다. 또한 사용자를 잘 파악하기 위해서 시간에 따라 관심의 변화를 파악하는 방법을 제시한다.

  • PDF

A Comparative Study on the Goodness of Fit in Spatial Econometric Models Using Housing Transaction Prices of Busan, Korea (부산시 실거래 주택매매 가격을 이용한 공간계량모형의 적합도 비교연구)

  • Chung, Kyoun-Sup;Kim, Sung-Woo;Lee, Yang-Won
    • Journal of the Korean Association of Geographic Information Studies
    • /
    • v.15 no.1
    • /
    • pp.43-51
    • /
    • 2012
  • The OLS(ordinary least squares) method is widely used in hedonic housing models. One of the assumptions of the OLS is an independent and uniform distribution of the disturbance term. This assumption can be violated when the spatial autocorrelation exists, which in turn leads to undesirable estimate results. An alterative to this, spatial econometric models have been introduced in housing price studies. This paper describes the comparisons between OLS and spatial econometric models using housing transaction prices of Busan, Korea. Owing to the approaches reflecting spatial autocorrelation, the spatial econometric models showed some superiority to the traditional OLS in terms of log likelihood and sigma square(${\sigma}^2$). Among the spatial models, the SAR(Spatial Autoregressive Models) seemed more appropriate than the SAC(General Spatial Models) and the SEM(Spatial Errors Models) for Busan housing markets. We can make sure the spatial effects on housing prices, and the reconstruction plans have strong impacts on the transaction prices. Selecting a suitable spatial model will play an important role in the housing policy of the government.