• Title/Summary/Keyword: data semantics

Search Result 252, Processing Time 0.026 seconds

Analyzing the Sentence Structure for Automatic Identification of Metadata Elements based on the Logical Semantic Structure of Research Articles (연구 논문의 의미 구조 기반 메타데이터 항목의 자동 식별 처리를 위한 문장 구조 분석)

  • Song, Min-Sun
    • Journal of the Korean Society for information Management
    • /
    • v.35 no.3
    • /
    • pp.101-121
    • /
    • 2018
  • This study proposes the analysis method in sentence semantics that can be automatically identified and processed as appropriate items in the system according to the composition of the sentences contained in the data corresponding to the logical semantic structure metadata of the research papers. In order to achieve the purpose, the structure of sentences corresponding to 'Research Objectives' and 'Research Outcomes' among the semantic structure metadata was analyzed based on the number of words, the link word types, the role of many-appeared words in sentences, and the end types of a word. As a result of this study, the number of words in the sentences was 38 in 'Research Objectives' and 212 in 'Research Outcomes'. The link word types in 'Research Objectives' were occurred in the order such as Causality, Sequence, Equivalence, In-other-word/Summary relation, and the link word types in 'Research Outcomes' were appeared in the order such as Causality, Equivalence, Sequence, In-other-word/Summary relation. Analysis target words like '역할(Role)', '요인(Factor)' and '관계(Relation)' played a similar role in both purpose and result part, but the role of '연구(Study)' was little different. Finally, the verb endings in sentences were appeared many times such as '~고자', '~였다' in 'Research Objectives', and '~었다', '~있다', '~였다' in 'Research Outcomes'. This study is significant as a fundamental research that can be utilized to automatically identify and input the metadata element reflecting the common logical semantics of research papers in order to support researchers' scholarly sensemaking.

Development of Subsurface Spatial Information Model with Cluster Analysis and Ontology Model (온톨로지와 군집분석을 이용한 지하공간 정보모델 개발)

  • Lee, Sang-Hoon
    • Journal of the Korean Association of Geographic Information Studies
    • /
    • v.13 no.4
    • /
    • pp.170-180
    • /
    • 2010
  • With development of the earth's subsurface space, the need for a reliable subsurface spatial model such as a cross-section, boring log is increasing. However, the ground mass was essentially uncertain. To generate model was uncertain because of the shortage of data and the absence of geotechnical interpretation standard(non-statistical uncertainty) as well as field environment variables(statistical uncertainty). Therefore, the current interpretation of the data and the generation of the model were accomplished by a highly trained experts. In this study, a geotechnical ontology model was developed using the current expert experience and knowledge, and the information content was calculated in the ontology hierarchy. After the relative distance between the information contents in the ontology model was combined with the distance between cluster centers, a cluster analysis that considered the geotechnical semantics was performed. In a comparative test of the proposed method, k-means method, and expert's interpretation, the proposed method is most similar to expert's interpretation, and can be 3D-GIS visualization through easily handling massive data. We expect that the proposed method is able to generate the more reasonable subsurface spatial information model without geotechnical experts' help.

Detection of Video Scene Boundaries based on the Local and Global Context Information (지역 컨텍스트 및 전역 컨텍스트 정보를 이용한 비디오 장면 경계 검출)

  • 강행봉
    • Journal of KIISE:Computing Practices and Letters
    • /
    • v.8 no.6
    • /
    • pp.778-786
    • /
    • 2002
  • Scene boundary detection is important in the understanding of semantic structure from video data. However, it is more difficult than shot change detection because scene boundary detection needs to understand semantics in video data well. In this paper, we propose a new approach to scene segmentation using contextual information in video data. The contextual information is divided into two categories: local and global contextual information. The local contextual information refers to the foreground regions' information, background and shot activity. The global contextual information refers to the video shot's environment or its relationship with other video shots. Coherence, interaction and the tempo of video shots are computed as global contextual information. Using the proposed contextual information, we detect scene boundaries. Our proposed approach consists of three consecutive steps: linking, verification, and adjusting. We experimented the proposed approach using TV dramas and movies. The detection accuracy of correct scene boundaries is over than 80%.

Automatic Generation of Machine Readable Context Annotations for SPARQL Results

  • Choi, Ji-Woong
    • Journal of the Korea Society of Computer and Information
    • /
    • v.21 no.10
    • /
    • pp.1-10
    • /
    • 2016
  • In this paper, we propose an approach to generate machine readable context annotations for SPARQL Results. According to W3C Recommendations, the retrieved data from RDF or OWL data sources are represented in tabular form, in which each cell's data is described by only type and value. The simple query result form is generally useful, but it is not sufficient to explain the semantics of the data in query results. To explain the meaning of the data, appropriate annotations must be added to the query results. In this paper, we generate the annotations from the basic graph patterns in user's queries. We could also manipulate the original queries to complete the annotations. The generated annotations are represented using the RDFa syntax in our study. The RDFa expressions in HTML are machine-understandable. We believe that our work will improve the trustworthiness of query results and contribute to distribute the data to meet the vision of the Semantic Web.

XML-BASED BIOINFORMATIC SYSTEMS (XML 기반의 생물정보학시스템)

  • Sin Jong Hyeon;Jeong Mu Yeong
    • Proceedings of the Korean Operations and Management Science Society Conference
    • /
    • 2002.05a
    • /
    • pp.301-305
    • /
    • 2002
  • Bioinformatics can beexplained as the science of developing algorithms, applicatiou tools, and computer databases and so on, for the purpose of supporting and enhancing biological research. Bioinformatic information systems (BIS) typically handle large data sets and the amount of the data goes up exponentially. Another impediment to easy extraction and retrieval of genomic data in BIS is the need to access different sites for similar information. Recently. there has been some attempts to integrate bioinformatics data in the World Wide Web (WWW) among the bioinformatics community by the internet computing technology. However, the work to integrate bioinformatics data on a universal platform has some problems because of the lack of standard, terminologies, semantics, and ontologies about bioinformatics. In this paper, an XML-based BIS architecture is proposed as an integrated BIS framework. The XML and related technologies allow the creation of meaningful information tags to exchange data between various databases as a standard format, and to create more simple interfaces. This integrated BIS framework has bioinformatic architectural components which is used in the Corporate Information Factory (CIF) method.

  • PDF

A Study on the Method for Solving Data Heterogeneity in the Integrated Information System (통합 정보시스템에서의 데이터 이질성 해결 방안에 관한 연구)

  • Park, Seong-Jin;Park, Sung-Kong;Park, Hwa-Gyoo
    • Journal of Information Technology Services
    • /
    • v.7 no.4
    • /
    • pp.87-99
    • /
    • 2008
  • As the technologies for telecommunication have been evolving, more enhanced information services and integrated information systems have been introduced, which can manage a variety of information from the heterogeneous systems. The major obstacle for the integrated information systems is the integrating heterogeneous databases in the systems and the heterogeneity problems can be classified into the structural and data heterogeneities. However, the previous researches have mainly highlighted into the solving structural heterogeneity problems. This paper identifies the data heterogeneity problems for multi-database schema integrations and proposes a new solving method. We analyze the semantics equivalence in data values based on the functional dependency, primary and candidate keys, and present a procedural solution of data heterogeneity in the perspective of the concept of attribute equivalence, integration key and conceptual integration table.

Employing SNMP to Manage Ubiquitous Environments

  • Murtaza Syed Shariyar;Amin Syed Obaid;Hong Choong Seon;Choi Sang Hyun
    • Proceedings of the Korean Information Science Society Conference
    • /
    • 2005.11a
    • /
    • pp.196-198
    • /
    • 2005
  • The vision of ubiquitous computing is becoming a reality now .Service discovery in ubiquitous environment, as well as adding semantics to the data is quite established. But, not many researchers have thought about the management of these devices. We envisage that by integrating SNMP with service discovery protocols, we could leverage the use of management factors (like performance, fault and security management etc) in the business, organizations, and other multi-user environments.

  • PDF

Verification of Subsumption Anomalies in Hybrid Knowledge Bases : A Meta-graph Approach (혼합 지식 기반 내 포함 이상의 검증 메타 그라프적 접근)

  • Lee, Sun-Ro
    • Asia pacific journal of information systems
    • /
    • v.7 no.2
    • /
    • pp.163-183
    • /
    • 1997
  • As object models and hybrid knowledge are increasingly used in current information systems development, Is-a structures need to be more formally defined, and subsequently subsumption related anomalies need to be detected with minimal declaration of meta knowledge. This paper extends a metagraph in the hybrid environments and demonstrates its utilities for detecting such anomalies that can occur from semantics and dynamics unique to the hybrid knowledge and data structure.

  • PDF

Higher Order Knowledge Processing: Pathway Database and Ontologies

  • Fukuda, Ken Ichiro
    • Genomics & Informatics
    • /
    • v.3 no.2
    • /
    • pp.47-51
    • /
    • 2005
  • Molecular mechanisms of biological processes are typically represented as 'pathways' that have a graph­analogical network structure. However, due to the diversity of topics that pathways cover, their constituent biological entities are highly diverse and the semantics is embedded implicitly. The kinds of interactions that connect biological entities are likewise diverse. Consequently, how to model or process pathway data is not a trivial issue. In this review article, we give an overview of the challenges in pathway database development by taking the INOH project as an example.

A design and implementation of VHDL-to-C mapping in the VHDL compiler back-end (VHDL 컴파일러 후반부의 VHDL-to-C 사상에 관한 설계 및 구현)

  • 공진흥;고형일
    • Journal of the Korean Institute of Telematics and Electronics C
    • /
    • v.35C no.12
    • /
    • pp.1-12
    • /
    • 1998
  • In this paper, a design and implementation of VHDL-to-C mapping in the VHDL compiler back-end is described. The analyzed data in an intermediate format(IF), produced by the compiler front-end, is transformed into a C-code model of VHDL semantics by the VHDL-to-C mapper. The C-code model for VHDL semantics is based on a functional template, including declaration, elaboration, initialization and execution parts. The mapping is carried out by utilizing C mapping templates of 129 types classified by mapping units and functional semantics, and iterative algorithms, which are combined with terminal information, to produce C codes. In order to generate the C program, the C codes are output to the functional template either directly or by combining the higher mapping result with intermediate mapping codes in the data queue. In experiments, it is shown that the VHDL-to-C mapper could completely deal with the VHDL analyzed programs from the compiler front-end, which deal with about 96% of major VHDL syntactic programs in the Validation Suite. As for the performance, it is found that the code size of VHDL-to-C is less than that of interpreter and worse than direct code compiler of which generated code is increased more rapidly with the size of VHDL design, and that the VHDL-to-C timing overhead is needed to be improved by the optimized implementation of mapping mechanism.

  • PDF