• Title/Summary/Keyword: RDF Data

Search Result 195, Processing Time 0.027 seconds

A Study of Methodology for Automatic Construction of OWL Ontologies from Sejong Electronic Dictionary (대용량 OWL 온톨로지 자동구축을 위한 세종전자사전 활용 방법론 연구)

  • Song Do Gyu
    • Language and Information
    • /
    • v.9 no.1
    • /
    • pp.19-34
    • /
    • 2005
  • Ontology is an indispensable component in intelligent and semantic processing of knowledge and information, such as in semantic web. However, ontology construction requires vast amount of data collection and arduous efforts in processing these un-structured data. This study proposed a methodology to automatically construct and generate ontologies from Sejong Electronic Dictionary. As Sejong Electronic Dictionary is structured in XML format, it can be processed automatically by computer programmed tools into an OWL(Web Ontology Language)-based ontologies as specified in W3C . This paper presents the process and concrete application of this methodology.

  • PDF

An elastic distributed parallel Hadoop system for bigdata platform and distributed inference engines (동적 분산병렬 하둡시스템 및 분산추론기에 응용한 서버가상화 빅데이터 플랫폼)

  • Song, Dong Ho;Shin, Ji Ae;In, Yean Jin;Lee, Wan Gon;Lee, Kang Se
    • Journal of the Korean Data and Information Science Society
    • /
    • v.26 no.5
    • /
    • pp.1129-1139
    • /
    • 2015
  • Inference process generates additional triples from knowledge represented in RDF triples of semantic web technology. Tens of million of triples as an initial big data and the additionally inferred triples become a knowledge base for applications such as QA(question&answer) system. The inference engine requires more computing resources to process the triples generated while inferencing. The additional computing resources supplied by underlying resource pool in cloud computing can shorten the execution time. This paper addresses an algorithm to allocate the number of computing nodes "elastically" at runtime on Hadoop, depending on the size of knowledge data fed. The model proposed in this paper is composed of the layered architecture: the top layer for applications, the middle layer for distributed parallel inference engine to process the triples, and lower layer for elastic Hadoop and server visualization. System algorithms and test data are analyzed and discussed in this paper. The model hast the benefit that rich legacy Hadoop applications can be run faster on this system without any modification.

A Caching Mechanism for Knowledge Maps (지식 맵을 위한 캐슁 기법)

  • 정준원;민경섭;김형주
    • Journal of KIISE:Computing Practices and Letters
    • /
    • v.10 no.3
    • /
    • pp.282-291
    • /
    • 2004
  • There has been many researches in TopicMap and RDF which are approach to handle data efficiently with metadata. However, No researches has been performed to service and implement except for presentation and description. In this paper, We suggest the caching mechanism to support an efficient access of knowledgemap and practical knowledgemap service with implementation of TopicMap system. First, We propose a method to navigate Knowledgemap efficiently that includes advantage of former methods. Then, To transmit TopicMap efficiently, We suggest caching mechanism for knowledgemap. This method is that user will be able to navigate knowledgemap efficiently in the viewpoint of human, not application. Therefor the mechanism doesn't cash topics by logical or physical locality but clustering by information and characteristic value of TopicMap. Lastly, we suggest replace mechanism by using graph structure of TopicMap for efficiency of transmission.

A Study on the Development of Ontology Management Tool (온톨로지 저작 도구 개발에 관한 연구)

  • Kim, Won-Pil;Kim, Jeong-Ho
    • Journal of the Institute of Electronics Engineers of Korea CI
    • /
    • v.45 no.6
    • /
    • pp.187-193
    • /
    • 2008
  • Nowadays, the study on e semantic web has been actively progressing for processing the web data semantically. For actualizing the semantic web environment, the core task is to build the ontology that defines the concepts and relations between concepts about the all things. Many ontology languages such as OWL, RDF(S), DAML+OIL were developed for building the ontology. And the many ontology tools were also implemented based on them. Although, many language and tools were researched, the practical use of the ontology tools is limited to the experts and researchers about the ontology because of the difficulty of the vocabulary, weak understanding about the ontology theory and the difficulty of the use of the ontology tools. And there are no studies on the reuse of constructed huge ontology. Therefore, in our study we design and implement the OWL ontology management tool that both the ontology experts and general users who want to build the ontologies are able to construct the ontology easily In this paper, we introduce the main modules used in our tool and features of our tool.

SPARQL Query Processing in Distributed In-Memory System (분산 메모리 시스템에서의 SPARQL 질의 처리)

  • Jagvaral, Batselem;Lee, Wangon;Kim, Kang-Pil;Park, Young-Tack
    • Journal of KIISE
    • /
    • v.42 no.9
    • /
    • pp.1109-1116
    • /
    • 2015
  • In this paper, we propose a query processing approach that uses the Spark functional programming and distributed memory system to solve the computational overhead of SPARQL. In the semantic web, RDF ontology data is produced at large scale, and the main challenge for the semantic web is to query and manipulate such a large ontology with a high throughput. The most existing studies on SPARQL have focused on deploying the Hadoop MapReduce framework, and although approaches based on Hadoop MapReduce have shown promising results, they achieve a low level of throughput due to the underlying distributed file processes. Therefore, in order to speed up the query processes, we suggest query- processing methods that are based on memory caching in distributed memory system. Our approach is also integrated with a clause unification method for propagating between the clauses that exploits Spark join, map and filter methods along with caching. In our experiments, we have achieved a high level of performance relative to other approaches. In particular, our performance was nearly similar to that of Sempala, which has been considered to be the fastest query processing system.

DRAZ: SPARQL Query Engine for heterogeneous metadata sources (DRAZ : 이기종 메타 데이터 소스를 위한 SPARQL 쿼리 엔진)

  • Qudus, UMAIR;Hossain, Md Ibrahim;Lee, ChangJu;Khan, Kifayat Ullah;Won, Heesun;Lee, Young-Koo
    • Database Research
    • /
    • v.34 no.3
    • /
    • pp.69-85
    • /
    • 2018
  • Many researches proposed federated query engines to perform query on several homogeneous or heterogeneous datasets simultaneously that significantly improve the quality of query results. The existing techniques allow querying only over a few heterogeneous datasets considering the static binding using the non-standard query. However, we observe that a simultaneous system considering the integration of heterogeneous metadata standards can offer better opportunity to generalize the query over any homogeneous and heterogeneous datasets. In this paper, we propose a transparent federated engine (DRAZ) to query over multiple data sources using SPARQL. In our system, we first develop the ontology for a non-RDF metadata standard based on the metadata kernel dictionary elements, which are standardized by the metadata provider. For a given SPARQL query, we translate any triple pattern into an API call to access the dataset of corresponding non-RDF metadata standard. We convert the results of every API call to N-triples and summarize the final results considering all triple patterns. We evaluated our proposed DRAZ using modified Fedbench benchmark queries over heterogeneous metadata standards, such as DCAT and DOI. We observed that DRAZ can achieve 70 to 100 percent correctness of the results despite the unavailability of the JOIN operations.

Standardization Trends of Open Web Data (개방형 웹 데이터 표준화 동향)

  • Kim, Chang-su;Kim, Sung-han;Lee, Seung-yun;Jung, Hoe-kyung
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2013.10a
    • /
    • pp.836-838
    • /
    • 2013
  • In recent years, the future direction of information technology social computing, mobile computing, cloud computing. Web technology industries beyond IT convergence technology for the service side of the parameters has been developed. In particular, the rapid increase of data in a web-based open Web and the importance of the next generation Web technology is increasing. In this paper, the next generation Web technology, an open Web and the importance of increasing domestic and international standardization trends were studied.

  • PDF

Change Acceptable In-Depth Searching in LOD Cloud for Efficient Knowledge Expansion (효과적인 지식확장을 위한 LOD 클라우드에서의 변화수용적 심층검색)

  • Kim, Kwangmin;Sohn, Yonglak
    • Journal of Intelligence and Information Systems
    • /
    • v.24 no.2
    • /
    • pp.171-193
    • /
    • 2018
  • LOD(Linked Open Data) cloud is a practical implementation of semantic web. We suggested a new method that provides identity links conveniently in LOD cloud. It also allows changes in LOD to be reflected to searching results without any omissions. LOD provides detail descriptions of entities to public in RDF triple form. RDF triple is composed of subject, predicates, and objects and presents detail description for an entity. Links in LOD cloud, named identity links, are realized by asserting entities of different RDF triples to be identical. Currently, the identity link is provided with creating a link triple explicitly in which associates its subject and object with source and target entities. Link triples are appended to LOD. With identity links, a knowledge achieves from an LOD can be expanded with different knowledge from different LODs. The goal of LOD cloud is providing opportunity of knowledge expansion to users. Appending link triples to LOD, however, has serious difficulties in discovering identity links between entities one by one notwithstanding the enormous scale of LOD. Newly added entities cannot be reflected to searching results until identity links heading for them are serialized and published to LOD cloud. Instead of creating enormous identity links, we propose LOD to prepare its own link policy. The link policy specifies a set of target LODs to link and constraints necessary to discover identity links to entities on target LODs. On searching, it becomes possible to access newly added entities and reflect them to searching results without any omissions by referencing the link policies. Link policy specifies a set of predicate pairs for discovering identity between associated entities in source and target LODs. For the link policy specification, we have suggested a set of vocabularies that conform to RDFS and OWL. Identity between entities is evaluated in accordance with a similarity of the source and the target entities' objects which have been associated with the predicates' pair in the link policy. We implemented a system "Change Acceptable In-Depth Searching System(CAIDS)". With CAIDS, user's searching request starts from depth_0 LOD, i.e. surface searching. Referencing the link policies of LODs, CAIDS proceeds in-depth searching, next LODs of next depths. To supplement identity links derived from the link policies, CAIDS uses explicit link triples as well. Following the identity links, CAIDS's in-depth searching progresses. Content of an entity obtained from depth_0 LOD expands with the contents of entities of other LODs which have been discovered to be identical to depth_0 LOD entity. Expanding content of depth_0 LOD entity without user's cognition of such other LODs is the implementation of knowledge expansion. It is the goal of LOD cloud. The more identity links in LOD cloud, the wider content expansions in LOD cloud. We have suggested a new way to create identity links abundantly and supply them to LOD cloud. Experiments on CAIDS performed against DBpedia LODs of Korea, France, Italy, Spain, and Portugal. They present that CAIDS provides appropriate expansion ratio and inclusion ratio as long as degree of similarity between source and target objects is 0.8 ~ 0.9. Expansion ratio, for each depth, depicts the ratio of the entities discovered at the depth to the entities of depth_0 LOD. For each depth, inclusion ratio illustrates the ratio of the entities discovered only with explicit links to the entities discovered only with link policies. In cases of similarity degrees with under 0.8, expansion becomes excessive and thus contents become distorted. Similarity degree of 0.8 ~ 0.9 provides appropriate amount of RDF triples searched as well. Experiments have evaluated confidence degree of contents which have been expanded in accordance with in-depth searching. Confidence degree of content is directly coupled with identity ratio of an entity, which means the degree of identity to the entity of depth_0 LOD. Identity ratio of an entity is obtained by multiplying source LOD's confidence and source entity's identity ratio. By tracing the identity links in advance, LOD's confidence is evaluated in accordance with the amount of identity links incoming to the entities in the LOD. While evaluating the identity ratio, concept of identity agreement, which means that multiple identity links head to a common entity, has been considered. With the identity agreement concept, experimental results present that identity ratio decreases as depth deepens, but rebounds as the depth deepens more. For each entity, as the number of identity links increases, identity ratio rebounds early and reaches at 1 finally. We found out that more than 8 identity links for each entity would lead users to give their confidence to the contents expanded. Link policy based in-depth searching method, we proposed, is expected to contribute to abundant identity links provisions to LOD cloud.

Design and Development of a System for Mapping of Medical Standard Terminologies (표준 의학 용어체계의 매핑을 위한 시스템의 설계 및 개발)

  • Lee, In-Keun;Kim, Hwa-Sun;Cho, Hune
    • Journal of the Korean Institute of Intelligent Systems
    • /
    • v.21 no.2
    • /
    • pp.237-243
    • /
    • 2011
  • Various standard terminologies in medical field are composed individually to different structure. Therefore, information on crosswalking between the terminologies is needed to combine and use the terminologies. Lots of mapping tools have been developed and used to create the information. However, since those tools deal with specific terminologies, the information is restrictly created. To overcome this problem, some tools have been developed, which perform mapping tasks by composing various terminologies. However, the tools also have difficulty of composing automatically the terminologies because the terminologies have different structures. Therefore, in this paper, we propose a method for composing and using the terminologies in the developed mapping system with keeping the original structure of the terminologies. In the proposed method, additional terminologies could be added on the mapping system and used by making metadata involving information on location and structure of the terminologies. And the mapping system could cope flexibly with the changes of the structure or context of the terminologies. Moreover, various types of mapping information could be defined and created in the system because mapping data are constructed as triplets in ontology. Therefore, the mapping data can be transformed and distributed in different formats such as OWL, RDF, and Excel. Finally, we confirmed the usefulness of the mapping system based on the proposed method through the experiments about creating mapping data.

Construction of Framework for Metadata Integration Using Master Data Approach (마스터 데이터를 활용한 메타데이터 통합 프레임워크 구축)

  • Lee, Seungmin
    • Journal of Korean Library and Information Science Society
    • /
    • v.44 no.1
    • /
    • pp.201-225
    • /
    • 2013
  • In order to overcome the problems of current approaches to metadata interoperability based on element mapping, this research proposed Master Element Framework that is an alternative approach to metadata interoperability. It is an approach to integrate metadata elements that have the same value, instead of direct mapping between similar elements. Master Element Framework is constructed to semantically integrate metadata elements in a hierarchical order and to interconnect between heterogeneous metadata standards. This approach is expected to be an alternative approach to metadata interoperability.