• Title/Summary/Keyword: 키워드 기반 시맨틱 링크

Search Result 5, Processing Time 0.018 seconds

OntoFrame: Semantic Web-based Inference Service (OntoFrame: 시맨틱 웹 기반의 추론 서비스)

  • Lee, Mi-Kyoung;Jung, Han-Min;Sung, Won-Kyung
    • 한국IT서비스학회:학술대회논문집
    • /
    • 2008.11a
    • /
    • pp.349-352
    • /
    • 2008
  • 본 논문에서는 시맨틱 웹 기반의 학술 정보 분석 서비스 프레임워크인 OntoFrame에 대해 소개하고자 한다. 2005년부터 개발되기 시작한 OntoFrame은 매년 새로운 서비스와 기술로 확장되고 있으며 OntoFrame2008에서는 다중 키워드 기반의 검색 서비스 및 다중 개체 중심적 통합 검색기능을 제공한다. 본 서비스는 키워드의 개체를 판단한 후에 인력, 주제, 인력+주제에 해당하는 서비스 API를 호출하여 추론 서비스 페이지를 구성한다. 이때 시스템에서 자동으로 판단되는 개체의 모호함을 제거하기 위해서 사용자의 의도라고 판단되는 최적의 개체 조합 페이지뿐만 아니라 해당 키워드에서 나타날 수 있는 모든 개체 조합의 후보 페이지들을 제공해주어 시스템의 일방적인 추천 서비스의 단점을 없앴다. 그리고 서비스의 결과로 제공되는 페이지에서 링크를 통한 추가조건 검색도 제공해 주어 사용자의 검색 의도를 정확하게 파악하여 편리한 정보 획득을 도와주는 시스템으로 개발하고 있다. OntoFrame2008은 여러 가지 풍부한 분석 서비스를 제공하여 연구자들이 학술 정보 검색 과정에 많은 도움이 되는 추론 서비스를 제공하고 있다.

  • PDF

Semantic Search System using Ontology-based Inference (온톨로지기반 추론을 이용한 시맨틱 검색 시스템)

  • Ha Sang-Bum;Park Yong-Tack
    • Journal of KIISE:Software and Applications
    • /
    • v.32 no.3
    • /
    • pp.202-214
    • /
    • 2005
  • The semantic web is the web paradigm that represents not general link of documents but semantics and relation of document. In addition it enables software agents to understand semantics of documents. We propose a semantic search based on inference with ontologies, which has the following characteristics. First, our search engine enables retrieval using explicit ontologies to reason though a search keyword is different from that of documents. Second, although the concept of two ontologies does not match exactly, can be found out similar results from a rule based translator and ontological reasoning. Third, our approach enables search engine to increase accuracy and precision by using explicit ontologies to reason about meanings of documents rather than guessing meanings of documents just by keyword. Fourth, domain ontology enables users to use more detailed queries based on ontology-based automated query generator that has search area and accuracy similar to NLP. Fifth, it enables agents to do automated search not only documents with keyword but also user-preferable information and knowledge from ontologies. It can perform search more accurately than current retrieval systems which use query to databases or keyword matching. We demonstrate our system, which use ontologies and inference based on explicit ontologies, can perform better than keyword matching approach .

Keyword-based networked knowledge map expressing content relevance between knowledge (지식 간 내용적 연관성을 표현하는 키워드 기반 네트워크형 지식지도 개발)

  • Yoo, Keedong
    • Journal of Intelligence and Information Systems
    • /
    • v.24 no.3
    • /
    • pp.119-134
    • /
    • 2018
  • A knowledge map as the taxonomy used in a knowledge repository should be structured to support and supplement knowledge activities of users who sequentially inquire and select knowledge for problem solving. The conventional knowledge map with a hierarchical structure has the advantage of systematically sorting out types and status of the knowledge to be managed, however it is not only irrelevant to knowledge user's process of cognition and utilization, but also incapable of supporting user's activity of querying and extracting knowledge. This study suggests a methodology for constructing a networked knowledge map that can support and reinforce the referential navigation, searching and selecting related and chained knowledge in term of contents, between knowledge. Regarding a keyword as the semantic information between knowledge, this research's networked knowledge map can be constructed by aggregating each set of knowledge links in an automated manner. Since a keyword has the meaning of representing contents of a document, documents with common keywords have a similarity in content, and therefore the keyword-based document networks plays the role of a map expressing interactions between related knowledge. In order to examine the feasibility of the proposed methodology, 50 research papers were randomly selected, and an exemplified networked knowledge map between them with content relevance was implemented using common keywords.

A Collaborative Video Annotation and Browsing System using Linked Data (링크드 데이터를 이용한 협업적 비디오 어노테이션 및 브라우징 시스템)

  • Lee, Yeon-Ho;Oh, Kyeong-Jin;Sean, Vi-Sal;Jo, Geun-Sik
    • Journal of Intelligence and Information Systems
    • /
    • v.17 no.3
    • /
    • pp.203-219
    • /
    • 2011
  • Previously common users just want to watch the video contents without any specific requirements or purposes. However, in today's life while watching video user attempts to know and discover more about things that appear on the video. Therefore, the requirements for finding multimedia or browsing information of objects that users want, are spreading with the increasing use of multimedia such as videos which are not only available on the internet-capable devices such as computers but also on smart TV and smart phone. In order to meet the users. requirements, labor-intensive annotation of objects in video contents is inevitable. For this reason, many researchers have actively studied about methods of annotating the object that appear on the video. In keyword-based annotation related information of the object that appeared on the video content is immediately added and annotation data including all related information about the object must be individually managed. Users will have to directly input all related information to the object. Consequently, when a user browses for information that related to the object, user can only find and get limited resources that solely exists in annotated data. Also, in order to place annotation for objects user's huge workload is required. To cope with reducing user's workload and to minimize the work involved in annotation, in existing object-based annotation automatic annotation is being attempted using computer vision techniques like object detection, recognition and tracking. By using such computer vision techniques a wide variety of objects that appears on the video content must be all detected and recognized. But until now it is still a problem facing some difficulties which have to deal with automated annotation. To overcome these difficulties, we propose a system which consists of two modules. The first module is the annotation module that enables many annotators to collaboratively annotate the objects in the video content in order to access the semantic data using Linked Data. Annotation data managed by annotation server is represented using ontology so that the information can easily be shared and extended. Since annotation data does not include all the relevant information of the object, existing objects in Linked Data and objects that appear in the video content simply connect with each other to get all the related information of the object. In other words, annotation data which contains only URI and metadata like position, time and size are stored on the annotation sever. So when user needs other related information about the object, all of that information is retrieved from Linked Data through its relevant URI. The second module enables viewers to browse interesting information about the object using annotation data which is collaboratively generated by many users while watching video. With this system, through simple user interaction the query is automatically generated and all the related information is retrieved from Linked Data and finally all the additional information of the object is offered to the user. With this study, in the future of Semantic Web environment our proposed system is expected to establish a better video content service environment by offering users relevant information about the objects that appear on the screen of any internet-capable devices such as PC, smart TV or smart phone.

A Mobile Landmarks Guide : Outdoor Augmented Reality based on LOD and Contextual Device (모바일 랜드마크 가이드 : LOD와 문맥적 장치 기반의 실외 증강현실)

  • Zhao, Bi-Cheng;Rosli, Ahmad Nurzid;Jang, Chol-Hee;Lee, Kee-Sung;Jo, Geun-Sik
    • Journal of Intelligence and Information Systems
    • /
    • v.18 no.1
    • /
    • pp.1-21
    • /
    • 2012
  • In recent years, mobile phone has experienced an extremely fast evolution. It is equipped with high-quality color displays, high resolution cameras, and real-time accelerated 3D graphics. In addition, some other features are includes GPS sensor and Digital Compass, etc. This evolution advent significantly helps the application developers to use the power of smart-phones, to create a rich environment that offers a wide range of services and exciting possibilities. To date mobile AR in outdoor research there are many popular location-based AR services, such Layar and Wikitude. These systems have big limitation the AR contents hardly overlaid on the real target. Another research is context-based AR services using image recognition and tracking. The AR contents are precisely overlaid on the real target. But the real-time performance is restricted by the retrieval time and hardly implement in large scale area. In our work, we exploit to combine advantages of location-based AR with context-based AR. The system can easily find out surrounding landmarks first and then do the recognition and tracking with them. The proposed system mainly consists of two major parts-landmark browsing module and annotation module. In landmark browsing module, user can view an augmented virtual information (information media), such as text, picture and video on their smart-phone viewfinder, when they pointing out their smart-phone to a certain building or landmark. For this, landmark recognition technique is applied in this work. SURF point-based features are used in the matching process due to their robustness. To ensure the image retrieval and matching processes is fast enough for real time tracking, we exploit the contextual device (GPS and digital compass) information. This is necessary to select the nearest and pointed orientation landmarks from the database. The queried image is only matched with this selected data. Therefore, the speed for matching will be significantly increased. Secondly is the annotation module. Instead of viewing only the augmented information media, user can create virtual annotation based on linked data. Having to know a full knowledge about the landmark, are not necessary required. They can simply look for the appropriate topic by searching it with a keyword in linked data. With this, it helps the system to find out target URI in order to generate correct AR contents. On the other hand, in order to recognize target landmarks, images of selected building or landmark are captured from different angle and distance. This procedure looks like a similar processing of building a connection between the real building and the virtual information existed in the Linked Open Data. In our experiments, search range in the database is reduced by clustering images into groups according to their coordinates. A Grid-base clustering method and user location information are used to restrict the retrieval range. Comparing the existed research using cluster and GPS information the retrieval time is around 70~80ms. Experiment results show our approach the retrieval time reduces to around 18~20ms in average. Therefore the totally processing time is reduced from 490~540ms to 438~480ms. The performance improvement will be more obvious when the database growing. It demonstrates the proposed system is efficient and robust in many cases.