• Title/Summary/Keyword: 데이타 종속성

Search Result 40, Processing Time 0.025 seconds

Design of a Hybrid Data Value Predictor with Dynamic Classification Capability in Superscalar Processors (슈퍼스칼라 프로세서에서 동적 분류 능력을 갖는 혼합형 데이타 값 예측기의 설계)

  • Park, Hee-Ryong;Lee, Sang-Jeong
    • Journal of KIISE:Computer Systems and Theory
    • /
    • v.27 no.8
    • /
    • pp.741-751
    • /
    • 2000
  • To achieve high performance by exploiting instruction level parallelism aggressively in superscalar processors, it is necessary to overcome the limitation imposed by control dependences and data dependences which prevent instructions from executing parallel. Value prediction is a technique that breaks data dependences by predicting the outcome of an instruction and executes speculatively its data dependent instruction based on the predicted outcome. In this paper, a hybrid value prediction scheme with dynamic classification mechanism is proposed. We design a hybrid predictor by combining the last predictor, a stride predictor and a two-level predictor. The choice of a predictor for each instruction is determined by a dynamic classification mechanism. This makes each predictor utilized more efficiently than the hybrid predictor without dynamic classification mechanism. To show performance improvements of our scheme, we simulate the SPECint95 benchmark set by using execution-driven simulator. The results show that our scheme effect reduce of 45% hardware cost and 16% prediction accuracy improvements comparing with the conventional hybrid prediction scheme and two-level value prediction scheme.

  • PDF

Implementation of CORBA based Spatial Data Provider for Interoperability (상호운용을 지원하는 코바 기반 공간 데이터 제공자의 설계 및 구현)

  • Kim, Min-Seok;An, Kyoung-Hwan;Hong, Bong-Hee
    • Journal of Korea Spatial Information System Society
    • /
    • v.1 no.2 s.2
    • /
    • pp.33-46
    • /
    • 1999
  • In distributed computing platforms like CORBA, wrappers are used to integrate heterogeneous systems or databases. A spatial data provider is one of the wrappers because it provides clients with uniform access interfaces to diverse data sources. The individual implementation of spatial data providers for each of different data sources is not efficient because of redundant coding of the wrapper modules. This paper presents a new architecture of the spatial data provider which consists of two layered objects : independent wrapper components and dependent wrapper components. Independent wrapper components would be reused for implementing a new data provider for a new data source, which dependent wrapper components should be newly coded for every data source. This paper furthermore discussed the issues of implementing the representation of query results in the middleware. There are two methods of keeping query results in the middleware. One is to keep query results as non-CORBA objects and the other is to transform query results into CORBA objects. The evaluation of the above two methods shows that the cost of making CORBA objects is very expensive.

  • PDF

A Message Conversion System based on MDR for Resolving Metadata Heterogeneity (메타데이타 이질성 해결을 위한 MDR 기반의 메시지 변환 시스템)

  • 김진관;김중일;정동원;백두권
    • Journal of KIISE:Databases
    • /
    • v.31 no.3
    • /
    • pp.232-242
    • /
    • 2004
  • Metadata is a general notion of data about data to improve data sharing and exchanging by definitely describing meaning and representation of data. However, metadata has been created in various ways and It caused another kind of heterogeneity problem named metadata heterogeneity problem. Recently, the research on metadata gateway approach that allows metadata heterogeneity is being more actively progressed. However, the existing commercialized systems that have been implemented with the metadata gateway approach are dependent on a metadata schema. In this paper, we propose a message conversion system which separates the mapping information from the mapping rules between heterogeneous metadata schemas. The proposed system dynamically manages standardized data elements by applying ISO/IEC l1179. Therefore, the proposed system provides the set of standard data elements to create consistently metadata of new databases and provides a fundamental resolution to the metadata heterogeneity problem.

An Efficient Logging Scheme based on Lazy Release Consistent Model for Distributed Shared Memory System (잠금 해제 지연 일관성 모델을 기반으로 하는 분산 공유 메모리 시스템에서의 효과적인 로깅기법)

  • Park, Tae-Soon;Yeom, Heon-Yeong
    • Journal of KIISE:Computer Systems and Theory
    • /
    • v.27 no.2
    • /
    • pp.188-199
    • /
    • 2000
  • This paper presents an efficient stable logging scheme for the distributed shared memory system based on the lazy release consistent memory model. In the proposed scheme, inter-process dependency is traced and stable logging is performed when the dependency relation between processes actually happens. With the dependency tracking, the proposed scheme requires much less frequency of stable logging, comparing with the previous schemes in which stable logging is performed whenever any information transfer happens between processes. Also, in the proposed scheme, every data item accessed by a process is not logged, but only the access information is logged in the stable storage. For the recovery from a failure, the correct version of the accessed data items can be effectively traced by using the logged access information. As a result, the amount of logged information is also reduced.

  • PDF

A Hybrid Value Predictor using Static and Dynamic Classification in Superscalar Processors (슈퍼스칼라 프로세서에서 정적 및 동적 분류를 사용한 혼합형 결과 값 예측기)

  • 김주익;박홍준;조영일
    • Journal of KIISE:Computer Systems and Theory
    • /
    • v.30 no.10
    • /
    • pp.569-578
    • /
    • 2003
  • Data dependencies are one of major hurdles to limit ILP(Instruction Level Parallelism), so several related works have suggested that the limit imposed by data dependencies can be overcome to some extent with use of the data value prediction. Hybrid value predictor can obtain the high prediction accuracy using advantages of various predictors, but it has a defect that same instruction has overlapping entries in all predictor. In this paper, we propose a new hybrid value predictor which achieves high performance by using the information of static and dynamic classification. The proposed predictor can enhance the prediction accuracy and efficiently decrease the prediction table size of predictor, because it allocates each instruction into single best-suited predictor during the fetch stage by using the information of static classification. Also, it can enhance the prediction accuracy because it selects a best- suited prediction method for the “Unknown”pattern instructions by using the dynamic classification mechanism. Simulation results based on the SimpleScalar/PISA tool set and the SPECint95 benchmarks show the average correct prediction rate of 85.1% by using the static classification mechanism. Also, we achieve the average correction prediction rate of 87.6% by using static and dynamic classification mechanism.

Multivariate Exponentially Weighted Moving Average(EWMA) Process Control and Statistical Process Monitoring in the Process Industry (장치산업에서 다변량 EWMA 공정제어와 통계적 공정감시)

  • 김복만;최성운
    • Journal of Korean Society of Industrial and Systems Engineering
    • /
    • v.15 no.26
    • /
    • pp.119-124
    • /
    • 1992
  • 본 논문은 장치산업에서 적용되는 다변량 EWMA 공정제어와 통계적 공정감시 통합시스템을 제안한다. 본 논문에서 제안한 통합시스템은 자동공정제어(APC)의 예측, 조정기능과 통계적 정정감시(SPM)의 이상점 발견 및 제거등의 각각의 장점을 이용하였다. 기존의 다변량 EWMA연구는 데이타간의 독립성을 가정하였으나 본 논문은 데이타간의 종속적인 형태인 IMA(1,1)모델을 대상으로 하였다.

  • PDF

Modeling and Verification of Workflows with Various Parallel Dependencies (다양한 병행 종속성을 포함한 워크플로우 모델링 및 검증)

  • 정희택;이도헌
    • The Journal of Information Technology and Database
    • /
    • v.6 no.1
    • /
    • pp.59-72
    • /
    • 1999
  • A study on workflow system as an automated business processing system is done recently. However, it did not consider various dependencies between parallel tasks. Therefore, this paper proposes modeling and verification of workflows with various parallel dependencies. For this, firstly, we propose four dependencies to specify various parallel dependencies between tasks. They contain sequential starts, parallel starts, sequential commits, and parallel commits. Secondly, we suggest a method to specify various parallel dependencies on workflow graph. Thirdly, we propose a verification method to detect contradictions on workflow specifications.

  • PDF

n-Gram/2L: A Space and Time Efficient Two-Level n-Gram Inverted Index Structure (n-gram/2L: 공간 및 시간 효율적인 2단계 n-gram 역색인 구조)

  • Kim Min-Soo;Whang Kyu-Young;Lee Jae-Gil;Lee Min-Jae
    • Journal of KIISE:Databases
    • /
    • v.33 no.1
    • /
    • pp.12-31
    • /
    • 2006
  • The n-gram inverted index has two major advantages: language-neutral and error-tolerant. Due to these advantages, it has been widely used in information retrieval or in similar sequence matching for DNA and Protein databases. Nevertheless, the n-gram inverted index also has drawbacks: the size tends to be very large, and the performance of queries tends to be bad. In this paper, we propose the two-level n-gram inverted index (simply, the n-gram/2L index) that significantly reduces the size and improves the query performance while preserving the advantages of the n-gram inverted index. The proposed index eliminates the redundancy of the position information that exists in the n-gram inverted index. The proposed index is constructed in two steps: 1) extracting subsequences of length m from documents and 2) extracting n-grams from those subsequences. We formally prove that this two-step construction is identical to the relational normalization process that removes the redundancy caused by a non-trivial multivalued dependency. The n-gram/2L index has excellent properties: 1) it significantly reduces the size and improves the Performance compared with the n-gram inverted index with these improvements becoming more marked as the database size gets larger; 2) the query processing time increases only very slightly as the query length gets longer. Experimental results using databases of 1 GBytes show that the size of the n-gram/2L index is reduced by up to 1.9${\~}$2.7 times and, at the same time, the query performance is improved by up to 13.1 times compared with those of the n-gram inverted index.

An XML Structure Translation System using Schema Structure Data Mapping (스키마 구조 데이타 매핑을 이용한 XML 구조변환 시스템)

  • 송종철;김창수;정회경
    • Journal of KIISE:Computing Practices and Letters
    • /
    • v.10 no.5
    • /
    • pp.406-418
    • /
    • 2004
  • Last days, various kinds of applications and system were individually introduced into specific groups or enterprises by different objective without considering interoperability among those. However, the environment for data processing is changing rapidly in these days. And now the necessity is growing to integrate and couple applications and system in the process dimension for more flexible and quicker data processing on these application programs and system. When integrating these application programs or system, an integration based on XML is recommended as it is one of good methods which will the additional cost and satisfy the requirements of the integration. This is because the XML is not only device-independent data type which can be used any platform, but also it uses XSLT, the document conversion standard established by W3C, which allows easy data conversion from one to another type on occasion of demands. This paper studies a design and implementation of system to convert XML structure. This system shows the structure of source- side providing data and destination-side processing data with using XML schema that defines structural information of a XML document. And this system defines the structure relationship of desired form as mapping structural information and data. This system creates the XSLT document that defines conversion rule between two structures based information which is defined. The XSLT document which is created as described above will convert data to be appropriate to the structure of the destination- side. By implementing this system, it is able to apply a document into various kinds of structure without considering specific system or platform and it is able to construct XSLT document to which meaning of desired form can be given. This paper aims to offer a process conversion between documents and to improve interoperability and scalability, so that we can contribute to build XML document processing environment

A Hybrid Value Predictor using Speculative Update of the Predictor Table and Static Classification for the Pattern of Executed Instructions in Superscalar Processors (슈퍼스칼라 프로세서에서 예상 테이블의 모험적 갱신과 명령어 실행 유형의 정적 분류를 이용한 혼합형 결과값 예측기)

  • Park, Hong-Jun;Jo, Young-Il
    • Journal of KIISE:Computing Practices and Letters
    • /
    • v.8 no.1
    • /
    • pp.107-115
    • /
    • 2002
  • We propose a new hybrid value predictor which achieves high performance by combining several predictors. Because the proposed hybrid value predictor can update the prediction table speculatively, it efficiently reduces the number of mispredicted instructions due to stale data. Also, the proposed predictor can enhance the prediction accuracy and efficiently decrease the hardware cost of predictor, because it allocates instructions into the best-suited predictor during instruction fetch stage by using the information of static classification which is obtained from the profile-based compiler implementation. For the 16-issue superscalar processors, simulation results based on the SimpleScalar/PISA tool set show that we achieve the average prediction rates of 73% by using speculative update and the average prediction rates of 88% by adding static classification for the SPECint95 benchmark programs.