• Title/Summary/Keyword: Consistency 알고리즘

Search Result 152, Processing Time 0.026 seconds

A Replicated Data Consistency Mechanism based on write-through cache coherence protocol for TDX system (전전자 교환기 시스템에서 write-through 캐쉬 일관성 프로토콜을 이용한 중복 데이터 일관성 유지 방안)

  • 원병재
    • Proceedings of the Korea Society for Simulation Conference
    • /
    • 1998.10a
    • /
    • pp.161-165
    • /
    • 1998
  • 다중 프로세서 구조로 실시간 분산 처리를 하는 전전자 교환기 시스템은 그 특성상 2개 이상의 프로세서에 동일한 값을 유지하는 중복 데이터의 사용이 필수적이다. 시스템의 자원 정보, 번호 번역 정보, 과금 정보 등이 중복 데이터로 사용된다. 이러한 중복 데이터에 대한 변경은 불일치 상태를 회피하기 위해 그 처리에 많은 비용과 제한이 따른다. 과도한 시그널 전송 및 로그 저장, 재전송 알고리즘은 데이터베이스 시스템의 성능을 저하시키고 때때로 순간적인 마비 상태까지도 유발할 수 있다. 본 논문에서는 기존 일관성 방안의 문제점을 분석하고 단일-버스 다중-프로세서 시스템에서 각각의 캐쉬들간의 일관성 유지를 위한 write-through 캐쉬 일관성 프로토콜을 사용하여 저 비용이며 효율적인 중복 데이터 일관성 유지 방안을 제시한다.

  • PDF

Design and Implementation of Client Spatial Views to Support Consistency Control (일관성 제어를 지원하는 클라이언트 공간뷰의 설계 및 구현)

  • 임덕성;반재훈;문상호;홍봉희
    • Proceedings of the Korean Information Science Society Conference
    • /
    • 1999.10a
    • /
    • pp.213-215
    • /
    • 1999
  • 클라이언트/서버 환경에서 서버의 공간 데이터를 검색, 접근하기 위한 공간 질의는 대용량의 공간 객체에 대해 복잡한 공간 연산을 수행한다. 그러므로 클라이언트가 자주 이용하는 공간 질의를 매번 서버에서 처리하는 경우에 서버의 부하가 증가하며 질의 응답시간도 길어지게 된다. 따라서 이러한 공간 질의를 뷰로 정의하고 클라이언트에 실체화하면 질의를 효율적으로 처리할 수 있다. 이 경우에 공간뷰를 유도한 서버의 소스 객체의 변경에 따라 클라이언트의 실체화된 뷰 객체의 일관성을 유지해야 한다. 본 논문에서는 클라이언트/서버 환경에서 공간뷰 개념을 확장한 클라이언트 공간뷰를 정의하고 일관성 유지를 위한 알고리즘을 제시한다. 그리고 상용 지리정보시스템인 고딕에서 클라이언트 공간뷰 시스템을 설계 및 구현한다. 마지막으로 질의 재수행과 본 논문에서 제시한 클라이언트 공간뷰의 성능을 실제 데이터를 이용하여 비교 평가한다.

  • PDF

Development of Chicken Carcass Segmentation Algorithm using Image Processing System (영상처리 시스템을 이용한 닭 도체 부위 분할 알고리즘 개발)

  • Cho, Sung-Ho;Lee, Hyo-Jai;Hwang, Jung-Ho;Choi, Sun;Lee, Hoyoung
    • Journal of the Korea Academia-Industrial cooperation Society
    • /
    • v.22 no.3
    • /
    • pp.446-452
    • /
    • 2021
  • As a higher standard for food consumption is required, the consumption of chicken meat that can satisfy the subdivided food preferences is increasing. In March 2003, the quality criteria for chicken carcasses notified by the Livestock Quality Assessment Service suggested quality grades according to fecal contamination and the size and weight of blood and bruises. On the other hand, it is too difficult for human inspection to qualify mass products, which is key to maintaining consistency for grading thousands of chicken carcasses. This paper proposed the computer vision algorithm as a non-destructive inspection, which can identify chicken carcass parts according to the detailed standards. To inspect the chicken carcasses conveyed at high speed, the image calibration was involved in providing robustness to the side effect of external lighting interference. The separation between chicken and background was achieved by a series of image processing, such as binarization based on Expectation Maximization, Erosion, and Labeling. In terms of shape analysis of chicken carcasses, the features are presented to reveal geometric information. After applying the algorithm to 78 chicken carcass samples, the algorithm was effective in segmenting chicken carcass against a background and analyzing its geometric features.

Image registration using outlier removal and triangulation-based local transformation (이상치 제거와 삼각망 기반의 지역 변환을 이용한 영상 등록)

  • Ye, Chul-Soo
    • Korean Journal of Remote Sensing
    • /
    • v.30 no.6
    • /
    • pp.787-795
    • /
    • 2014
  • This paper presents an image registration using Triangulation-based Local Transformation (TLT) applied to the remaining matched points after elimination of the matched points with gross error. The corners extracted using geometric mean-based corner detector are matched using Pearson's correlation coefficient and then accepted as initial matched points only when they satisfy the Left-Right Consistency (LRC) check. We finally accept the remaining matched points whose RANdom SAmple Consensus (RANSAC)-based global transformation (RGT) errors are smaller than a predefined outlier threshold. After Delaunay triangulated irregular networks (TINs) are created using the final matched points on reference and sensed images, respectively, affine transformation is applied to every corresponding triangle and then all the inner pixels of the triangles on the sensed image are transformed to the reference image coordinate. The proposed algorithm was tested using KOMPSAT-2 images and the results showed higher image registration accuracy than the RANSAC-based global transformation.

Efficient Deferred Incremental Refresh of XML Query Cache Using ORDBMS (ORDBMS를 사용한 XML 질의 캐쉬의 효율적인 지연 갱신)

  • Hwang Dae-Hyun;Kang Hyun-Chul
    • The KIPS Transactions:PartD
    • /
    • v.13D no.1 s.104
    • /
    • pp.11-22
    • /
    • 2006
  • As we are to deal with more and more XML documents, research on storing and managing XML documents in databases are actively conducted. Employing RDBMS or ORDBMS as a repository of XML documents is currently regarded as most practical. The query results out of XML documents stored in databases could be cached for query performance though the cost of cache consistency against the update of the underlying data is incurred. In this paper, we assume that an ORDBMS is used as a repository for the XML query cache as well as its underlying XML documents, and that XML query cache is refreshed in a deferred way with the update log. When the same XML document was updated multiple times, the deferred refresh of the XML query cache may Bet inefficient. We propose an algorithm that removes or filters such duplicate updates. Based on that, the optimal SQL statements that are to be executed for XML query cache consistency are generated. Through experiments, we show the efficiency of our proposed deferred refresh of XML query cache.

An Automated Topic Specific Web Crawler Calculating Degree of Relevance (연관도를 계산하는 자동화된 주제 기반 웹 수집기)

  • Seo Hae-Sung;Choi Young-Soo;Choi Kyung-Hee;Jung Gi-Hyun;Noh Sang-Uk
    • Journal of Internet Computing and Services
    • /
    • v.7 no.3
    • /
    • pp.155-167
    • /
    • 2006
  • It is desirable if users surfing on the Internet could find Web pages related to their interests as closely as possible. Toward this ends, this paper presents a topic specific Web crawler computing the degree of relevance. collecting a cluster of pages given a specific topic, and refining the preliminary set of related web pages using term frequency/document frequency, entropy, and compiled rules. In the experiments, we tested our topic specific crawler in terms of the accuracy of its classification, crawling efficiency, and crawling consistency. First, the classification accuracy using the set of rules compiled by CN2 was the best, among those of C4.5 and back propagation learning algorithms. Second, we measured the classification efficiency to determine the best threshold value affecting the degree of relevance. In the third experiment, the consistency of our topic specific crawler was measured in terms of the number of the resulting URLs overlapped with different starting URLs. The experimental results imply that our topic specific crawler was fairly consistent, regardless of the starting URLs randomly chosen.

  • PDF

A Method based on Ontology for detecting errors in the Software Design (온톨로지 기반의 소프트웨어 설계에러검출방법)

  • Seo, Jin-Won;Kim, Young-Tae;Kong, Heon-Tag;Lim, Jae-Hyun;Kim, Chi-Su
    • Journal of the Korea Academia-Industrial cooperation Society
    • /
    • v.10 no.10
    • /
    • pp.2676-2683
    • /
    • 2009
  • The objective of this thesis is to improve the quality of a software product based on the enhancement of a software design quality using a better error detecting method. Also, this thesis is based on a software design method called as MOA(Methodology for Object to Agents) which uses an ontology based ODES(A Method based on Ontology for Detecting Errors in the Software Design) model as a common information model. At this thesis, a new format of error detecting method was defined. The method is implemented during a transformation process from UML model to ODES model using a ODES model, a Inter-View Inconsistency Detection technique and a combination of ontologic property of consistency framework and related rules. Transformation process to ODES model includes lexicon analysis and meaning analysis of a software design using of multiple mapping table at algorithm for the generation of ODES model instance.

Data Update on Multi-Scale Databases (다중축척 공간 데이터베이스의 데이터 갱신)

  • Kwon O-Je;Kang Hae-Kyong;Li Ki-Joune
    • Spatial Information Research
    • /
    • v.12 no.3
    • /
    • pp.239-249
    • /
    • 2004
  • This paper discusses on the update problem of multi-scale databases when the multi-scale databases, which is several spatial databases covering the same geographic area with different scales, are derived from an original one. Although the integrity between original and derived multi-scale databases should be maintained, most of update mechanisms do not 6respect it since the update mechanisms have assumed that the update of source objects propagates to objects directly derived from the source. In order to maintain the integrity of multi-scale databases during updates, we must propagate updates of sources to objects derived from both the updated source objects and other related objects. It is an important functional requirement of multi-scale database systems, which has not been supported by existing spatial database systems. In this paper, we propose a set of rules and algorithms for the update propagation and show a prototype developed on ArcGIS of ESRI. Our update mechanism provides with not only the consistency between multi-scale databases but also incremental updates.

  • PDF

An Extensible Transaction Model for Real-Time Data Processing (실시간 데이타 처리를 위한 확장 가능한 트랜잭션 모델에 관한 연구)

  • 문승진
    • Journal of Internet Computing and Services
    • /
    • v.1 no.2
    • /
    • pp.11-18
    • /
    • 2000
  • In this paper we present a new extensible model based upon the concept of subtransactions in real-time transaction systems. The nested transaction model originally proposed by J. Moss is extended for real-time uniprocessor transaction systems by adding explicit timing constraints. Based upon the model, an integrated concurrency control and scheduling algorithm is developed, that not only guarantees timing constraints of a set of real-time transactions but also maintains consistency of the database. The algorithm is based on the priority ceiling protocol of Sha et al. We prove that the Real-Time Nested Priority Ceiling Protocol prevents unbounded blocking and deadlock, and maintains the serializability of a set of real-time transactions. We use the upper bound on the duration that a transaction can be blocked to show that it is possible to analyze the schedulability of a transaction set using rate-monotonic priority assignment. This work is viewed as a step toward multiprocessor and distributed real-time nested transaction systems. Also, it is possible to be extended to include the real-time multimedia transactions in the emerging web-based database application areas.

  • PDF

Efficient Cooperative Caching Algorithm for Distributed File Systems (분산 파일시스템을 위한 효율적인 협력캐쉬 알고리즘)

  • 박새미;이석재;유재수
    • Proceedings of the Korea Contents Association Conference
    • /
    • 2003.11a
    • /
    • pp.234-244
    • /
    • 2003
  • In distributed file-systems, cooperative caching algorithm which owns the data cached at each node jointly is used to reduce an expense of disk access. Cooperative caching algorithm is the method that increases a cache hit-ratio and decrease a disk access as it holds the cache information of distributed systems in common and makes cache larger virtually. Recently, several cooperative caching algorithms decrease the message costs by using approximate information of the cache and increase the cache hit-ratio by using local and global cache fields dynamically. And they have an advantage that increases the whole field hit-ratio by sending a replaced block to the idel node on cache replacement in order to maintain the replaced block in the cache field. However the wrong approximate information deteriorates the performance, the concistency maintenance goes to great expense to exchange messeges and the cost that manages Age-information of each node to choose the idle node increases. In this thesis, we propose a cooperative cache algorithm that maintains correct cache information, minimizes the maintance cost for consistency and the management cost for cache Age-information. Also, we show the superiority of our algorithm through the performance evaluation.

  • PDF