• Title/Summary/Keyword: incremental update

Search Result 58, Processing Time 0.031 seconds

Storage Schemes for XML Query Cache (XML 질의 캐쉬의 저장 기법)

  • Kim, Young-Hyun;Kang, Hyun-Chul
    • Journal of KIISE:Databases
    • /
    • v.33 no.5
    • /
    • pp.551-562
    • /
    • 2006
  • XML query caching for XML database-backed Web applications began to be investigated recently. Despite its practical significance, efficiency of the storage schemes for cached query results has not been addressed. In this paper, we deal with the storage schemes for XML query cache. A fundamental problem that needs to be considered in designing an efficient storage structure for XML query cache is that there exist performance tradeoffs between the two major types of operations on a cached query result. The two are (1) retrieving the whole of it to return the query result and (2) updating just a small portion of it for its incremental refresh against the updates done to its source. We propose eight different storage schemes for XML query cache, which are categorized into three groups: (1) the schemes based on the plain text file, (2) the schemes based on the persistent DOM (PDOM) file, and (3) a scheme employing an RDBMS. We implemented all of them, and compared their performance with each other. We also compared our proposal with a storage scheme based on a current state-of-the-art XML storage scheme, showing that ours is more efficient.

An Architecture and Experimentation for Efficient and Reliable Name Service Systems (효율적이고 신뢰성 있는 네임 서비스 시스템의 구조 및 실험)

  • 심영철;박준철;강호석;이준원
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.29 no.3B
    • /
    • pp.375-390
    • /
    • 2004
  • Domain Name System(DNS), one of the most important Internet services, handles mapping from host names to Internet addresses and vice versa, and precedes many Internet applications such as Web, e-mail, file transfer, etc. In this paper, we propose a structural design of a generic name server system providing name services for a huge domain for the purpose of improving the performance as well as the reliability of the system. We demonstrate the validity of the design by implementing and running a testbed system. Our testbed employs a couple of master name sowers for distributing the service overhead over two, rather than one, servers and for achieving high availability of the system as a whole. We suggest the use of dynamic update to add and delete records from a zone for which the name server has authority. The slave name servers located remotely then get a new, updated copy of the zone via incremental zone transfers(IXFRs). The experiments with the implemented testbed show that the proposed structure would easily manage increasing demands on the server power, and be highly available in the face of transient faults of a module in the system.

An Efficient Incremental View Maintenance in Data Warehouses (데이타 웨어하우스에서 효과적인 점진적 뷰 관리)

  • Lee, Ki-Yong;Kim, Myoung-Ho
    • Journal of KIISE:Databases
    • /
    • v.27 no.2
    • /
    • pp.175-184
    • /
    • 2000
  • A data warehouse is an integrated and summarized collection of data that can efficiently support decision making process. The summarized data at the data warehouse is often stored in materialized views. These materialized views need to be updated when source data change. Since the propagation of updates to the views may impose a significant overhead, it is very important to update the warehouse views efficiently. Though various strategies have been proposed to maintain views in the past, they typically require too much accesses to the data sources when the changes of multiple data sources have to be reflected in the view. In this paper we propose an efficient view update strategy that uses relatively small number of accesses to the data sources. We also show the performance advantage of our method over other existing methods through experiments using TPC-D data and queries.

  • PDF

A Study on Data Caching and Updates for Efficient Spatial Query Processing in Client/Server Environments (클라이언트/서버 환경에서 효율적인 공간질의 처리를 위한 데이터 캐싱과 변경에 관한 연구)

  • 문상호
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.7 no.6
    • /
    • pp.1269-1275
    • /
    • 2003
  • This paper addresses several issues on data caching and consistency of cached data in order to process client's queries efficiently in client/server environments. For the purpose, first of all, materialized spatial views are adapted in a client side for data caching, which is called client views. Also, an incremental update scheme using derivation relationships is applied to keep cached data of clients consistent with the rest of server databases. Materialized views support efficient query processing in a client side, however, it is difficult to keep consistent their contents by the update of a server database. In this paper, we devise cost functions on query execution and view maintenance based the cost of spatial operators so as to process client's queries efficiently. When request the client's query, in our query processing scheme, the server determines whether or not materialize it as a view due to evaluation using the related cost functions. Since the scheme supports a hybrid approach based on both view materialization and re-execution, hence, it should improve query execution times in client/server environments.

Managing the Reverse Extrapolation Model of Radar Threats Based Upon an Incremental Machine Learning Technique (점진적 기계학습 기반의 레이더 위협체 역추정 모델 생성 및 갱신)

  • Kim, Chulpyo;Noh, Sanguk
    • The Journal of Korean Institute of Next Generation Computing
    • /
    • v.13 no.4
    • /
    • pp.29-39
    • /
    • 2017
  • Various electronic warfare situations drive the need to develop an integrated electronic warfare simulator that can perform electronic warfare modeling and simulation on radar threats. In this paper, we analyze the components of a simulation system to reversely model the radar threats that emit electromagnetic signals based on the parameters of the electronic information, and propose a method to gradually maintain the reverse extrapolation model of RF threats. In the experiment, we will evaluate the effectiveness of the incremental model update and also assess the integration method of reverse extrapolation models. The individual model of RF threats are constructed by using decision tree, naive Bayesian classifier, artificial neural network, and clustering algorithms through Euclidean distance and cosine similarity measurement, respectively. Experimental results show that the accuracy of reverse extrapolation models improves, while the size of the threat sample increases. In addition, we use voting, weighted voting, and the Dempster-Shafer algorithm to integrate the results of the five different models of RF threats. As a result, the final decision of reverse extrapolation through the Dempster-Shafer algorithm shows the best performance in its accuracy.

A Comparison of Data Extraction Techniques and an Implementation of Data Extraction Technique using Index DB -S Bank Case- (원천 시스템 환경을 고려한 데이터 추출 방식의 비교 및 Index DB를 이용한 추출 방식의 구현 -ㅅ 은행 사례를 중심으로-)

  • 김기운
    • Korean Management Science Review
    • /
    • v.20 no.2
    • /
    • pp.1-16
    • /
    • 2003
  • Previous research on data extraction and integration for data warehousing has concentrated mainly on the relational DBMS or partly on the object-oriented DBMS. Mostly, it describes issues related with the change data (deltas) capture and the incremental update by using the triggering technique of active database systems. But, little attention has been paid to data extraction approaches from other types of source systems like hierarchical DBMS, etc. and from source systems without triggering capability. This paper argues, from the practical point of view, that we need to consider not only the types of information sources and capabilities of ETT tools but also other factors of source systems such as operational characteristics (i.e., whether they support DBMS log, user log or no log, timestamp), and DBMS characteristics (i.e., whether they have the triggering capability or not, etc), in order to find out appropriate data extraction techniques that could be applied to different source systems. Having applied several different data extraction techniques (e.g., DBMS log, user log, triggering, timestamp-based extraction, file comparison) to S bank's source systems (e.g., IMS, DB2, ORACLE, and SAM file), we discovered that data extraction techniques available in a commercial ETT tool do not completely support data extraction from the DBMS log of IMS system. For such IMS systems, a new date extraction technique is proposed which first creates Index database and then updates the data warehouse using the Index database. We illustrates this technique using an example application.

Adopting DITA Standard for Developing Technical Documentation Efficiently (효율적인 기술문서 개발을 위한 DITA 표준 적용)

  • Koo, Heung-Seo
    • The Transactions of The Korean Institute of Electrical Engineers
    • /
    • v.65 no.1
    • /
    • pp.171-178
    • /
    • 2016
  • In engineering industry, technical documentation refers to any type of documentation that describes handling, functionality and architecture of a technical product. The intended audiences for product technical documentation is the end user, the administrator, service or maintenance technician. Competition and rapid technological evolutions have created pressure to release new and improved products on an increasingly frequent basis. As the technical capabilities of products advance, the technical documentation for those products is becoming longer and complex. The Darwin Information Typing Architecture (DITA) is and XML-based, end-to-end architecture for authoring, producing, and delivering technical documents. DITA is an OASIS standard that allows documentation groups to single-source document for multiple products and users, to automatically publish that documentation in several media formats including PDF and HTML, and to efficiently maintain and update that documentation. This paper study on the various implementation projects for technical documentation using OASIS DITA standard and examines the potential for using DITA as a solution for reconstructing existing technical documentation. Also We offer an incremental adoption approach to effectively implement DITA standard of technical documentation project for domestic firms, while reducing failures.

Temporal Database Management Testbed (시간 지원 데이타 베이스 관리 시험대)

  • Kim, Dong-Ho;Jeon, Geun-Hwan
    • The Transactions of the Korea Information Processing Society
    • /
    • v.1 no.1
    • /
    • pp.1-13
    • /
    • 1994
  • The Temporal Database Management Testbed supports valid and transaction time. In this paper, we discuss the design and implementation of a testbed of a temporal database management system in main memory. The testbed consists of a syntactic analyzer, a semantic analyzer, a code generator, and an interpreter. The syntactic analyzer builds a parse tree from a temporal query. The semantic analyzer then checks it for correctness against the system catalog. The code generator builds an execution tree termed ann update network. We employ an incremental view materialization for the execution tree. After building the execution tree, the interpreter activates each node of the execution tree. Also, the indexing structure and the concurrency control are discussed in the testbed.

  • PDF

Preliminary study on the thermal-mechanical performance of the U3Si2/Al dispersion fuel plate under normal conditions

  • Yang, Guangliang;Liao, Hailong;Ding, Tao;Chen, Hongli
    • Nuclear Engineering and Technology
    • /
    • v.53 no.11
    • /
    • pp.3723-3740
    • /
    • 2021
  • The harsh conditions in the reactor affect the thermal and mechanical performance of the fuel plate heavily. Some in-pile behaviors, like fission-induced swelling, can cause a large deformation of fuel plate at very high burnup, which may even disturb the flow of coolant. In this research, the emphasis is put on the thermal expansion, fission-induced swelling, interaction layer (IL) growth, creep of the fuel meat, and plasticity of the cladding for the U3Si2/Al dispersion fuel plate. A detailed model of the fuel meat swelling is developed. Taking these in-pile behaviors into consideration, the three-dimensional large deformation incremental constitutive relations and stress update algorithms have been developed to study its thermal-mechanical performance under normal conditions using Abaqus. Results have shown that IL can effectively decrease the thermal conductivity of fuel meat. The high Mises stress region mainly locates at the interface between fuel meat and cladding, especially around the side edge of the interface. With irradiation time increasing, the stress in the fuel plate gets larger resulting from the growth of fuel meat swelling but then decreases under the effect of creep deformation. For the cladding, plasticity deformation does not occur within the irradiation time.

Improving Naïve Bayes Text Classifiers with Incremental Feature Weighting (점진적 특징 가중치 기법을 이용한 나이브 베이즈 문서분류기의 성능 개선)

  • Kim, Han-Joon;Chang, Jae-Young
    • The KIPS Transactions:PartB
    • /
    • v.15B no.5
    • /
    • pp.457-464
    • /
    • 2008
  • In the real-world operational environment, most of text classification systems have the problems of insufficient training documents and no prior knowledge of feature space. In this regard, $Na{\ddot{i}ve$ Bayes is known to be an appropriate algorithm of operational text classification since the classification model can be evolved easily by incrementally updating its pre-learned classification model and feature space. This paper proposes the improving technique of $Na{\ddot{i}ve$ Bayes classifier through feature weighting strategy. The basic idea is that parameter estimation of $Na{\ddot{i}ve$ Bayes considers the degree of feature importance as well as feature distribution. We can develop a more accurate classification model by incorporating feature weights into Naive Bayes learning algorithm, not performing a learning process with a reduced feature set. In addition, we have extended a conventional feature update algorithm for incremental feature weighting in a dynamic operational environment. To evaluate the proposed method, we perform the experiments using the various document collections, and show that the traditional $Na{\ddot{i}ve$ Bayes classifier can be significantly improved by the proposed technique.