• Title/Summary/Keyword: Semantic Knowledge-based Model

Search Result 114, Processing Time 0.029 seconds

A 3-Layered Information Integration System based on MDRs End Ontology (MDR과 온톨로지를 결합한 3계층 정보 통합 시스템)

  • Baik, Doo-Kwon;Choi, Yo-Han;Park, Sung-Kong;Lee, Jeong-Oog;Jeong, Dong-Won
    • The KIPS Transactions:PartD
    • /
    • v.10D no.2
    • /
    • pp.247-260
    • /
    • 2003
  • To share and standardize information, especially in the database environments, MDR (Metadata Registry) can be used to integrate various heterogeneous databases within a particular domain. But due to the discrepancies of data element representation between organizations, global information integration is not so easy. And users who are searching integrated information on the Web have limitation to obtain schema information for the underlying source databases. To solve those problems, in this paper, we present a 3-layered Information Integration System (LI2S) based on MDRs and Ontology. The purpose of proposed architecture is to define information integration model, which combine both of the nature of MDRs standard specification and functionality of ontology for the concept and relation. Adopting agent technology to the proposed model plays a key role to support the hierarchical and independent information integration architecture. Ontology is used as for a role of semantic network from which it extracts concept from the user query and the establishment of relationship between MDRs for the data element. (MDR and Knowledge Base are used as for the solution of discrepancies of data element representation between MDRs. Based on this architectural concept, LI2S was designed and implemented.

A Study on Knowledge Entity Extraction Method for Individual Stocks Based on Neural Tensor Network (뉴럴 텐서 네트워크 기반 주식 개별종목 지식개체명 추출 방법에 관한 연구)

  • Yang, Yunseok;Lee, Hyun Jun;Oh, Kyong Joo
    • Journal of Intelligence and Information Systems
    • /
    • v.25 no.2
    • /
    • pp.25-38
    • /
    • 2019
  • Selecting high-quality information that meets the interests and needs of users among the overflowing contents is becoming more important as the generation continues. In the flood of information, efforts to reflect the intention of the user in the search result better are being tried, rather than recognizing the information request as a simple string. Also, large IT companies such as Google and Microsoft focus on developing knowledge-based technologies including search engines which provide users with satisfaction and convenience. Especially, the finance is one of the fields expected to have the usefulness and potential of text data analysis because it's constantly generating new information, and the earlier the information is, the more valuable it is. Automatic knowledge extraction can be effective in areas where information flow is vast, such as financial sector, and new information continues to emerge. However, there are several practical difficulties faced by automatic knowledge extraction. First, there are difficulties in making corpus from different fields with same algorithm, and it is difficult to extract good quality triple. Second, it becomes more difficult to produce labeled text data by people if the extent and scope of knowledge increases and patterns are constantly updated. Third, performance evaluation is difficult due to the characteristics of unsupervised learning. Finally, problem definition for automatic knowledge extraction is not easy because of ambiguous conceptual characteristics of knowledge. So, in order to overcome limits described above and improve the semantic performance of stock-related information searching, this study attempts to extract the knowledge entity by using neural tensor network and evaluate the performance of them. Different from other references, the purpose of this study is to extract knowledge entity which is related to individual stock items. Various but relatively simple data processing methods are applied in the presented model to solve the problems of previous researches and to enhance the effectiveness of the model. From these processes, this study has the following three significances. First, A practical and simple automatic knowledge extraction method that can be applied. Second, the possibility of performance evaluation is presented through simple problem definition. Finally, the expressiveness of the knowledge increased by generating input data on a sentence basis without complex morphological analysis. The results of the empirical analysis and objective performance evaluation method are also presented. The empirical study to confirm the usefulness of the presented model, experts' reports about individual 30 stocks which are top 30 items based on frequency of publication from May 30, 2017 to May 21, 2018 are used. the total number of reports are 5,600, and 3,074 reports, which accounts about 55% of the total, is designated as a training set, and other 45% of reports are designated as a testing set. Before constructing the model, all reports of a training set are classified by stocks, and their entities are extracted using named entity recognition tool which is the KKMA. for each stocks, top 100 entities based on appearance frequency are selected, and become vectorized using one-hot encoding. After that, by using neural tensor network, the same number of score functions as stocks are trained. Thus, if a new entity from a testing set appears, we can try to calculate the score by putting it into every single score function, and the stock of the function with the highest score is predicted as the related item with the entity. To evaluate presented models, we confirm prediction power and determining whether the score functions are well constructed by calculating hit ratio for all reports of testing set. As a result of the empirical study, the presented model shows 69.3% hit accuracy for testing set which consists of 2,526 reports. this hit ratio is meaningfully high despite of some constraints for conducting research. Looking at the prediction performance of the model for each stocks, only 3 stocks, which are LG ELECTRONICS, KiaMtr, and Mando, show extremely low performance than average. this result maybe due to the interference effect with other similar items and generation of new knowledge. In this paper, we propose a methodology to find out key entities or their combinations which are necessary to search related information in accordance with the user's investment intention. Graph data is generated by using only the named entity recognition tool and applied to the neural tensor network without learning corpus or word vectors for the field. From the empirical test, we confirm the effectiveness of the presented model as described above. However, there also exist some limits and things to complement. Representatively, the phenomenon that the model performance is especially bad for only some stocks shows the need for further researches. Finally, through the empirical study, we confirmed that the learning method presented in this study can be used for the purpose of matching the new text information semantically with the related stocks.

Component Grid: A Developer-centric Environment for Defense Software Reuse (컴포넌트 그리드: 개발자 친화적인 국방 소프트웨어 재사용 지원 환경)

  • Ko, In-Young;Koo, Hyung-Min
    • Journal of Software Engineering Society
    • /
    • v.23 no.4
    • /
    • pp.151-163
    • /
    • 2010
  • In the defense software domain where large-scale software products in various application areas need to be built, reusing software is regarded as one of the important practices to build software products efficiently and economically. There have been many efforts to apply various methods to support software reuse in the defense software domain. However, developers in the defense software domain still experience many difficulties and face obstacles in reusing software assets. In this paper, we analyze practical problems of software reuse in the defense software domain, and define core requirements to solve those problems. To meet these requirements, we are currently developing the Component Grid system, a reuse-support system that provides a developer-centric software reuse environment. We have designed an architecture of Component Grid, and defined essential elements of the architecture. We have also developed the core approaches for developing the Component Grid system: a semantic-tagging-based requirement tracing method, a reuse-knowledge representation model, a social-network-based asset search method, a web-based asset management environment, and a wiki-based collaborative and participative knowledge construction and refinement method. We expect that the Component Grid system will contribute to increase the reusability of software assets in the defense software domain by providing the environment that supports transparent and efficient sharing and reuse of software assets.

  • PDF

A Study of Holism based Service Experience Analysis System

  • Kim, Sung-Su;Lee, Eun-Jong
    • Journal of the Ergonomics Society of Korea
    • /
    • v.31 no.1
    • /
    • pp.49-61
    • /
    • 2012
  • Objective: The aim of this study is to offer a holism based Service Experience Analysis System(HSEAS) for service design. Background: Customer experience has been focused in a lot of service area. But it is difficult to understand the customer's needs and their experiences because it's so complex and uncertain. Thus it needs holistic approach that means it's difficult to explain general character merely through the understanding of small parts that composes an object and it must be analyzed within the overall context. Method: Accordingly, the thesis paper proposes the Service Experience Analysis System that satisfies the four following needs. (1) Need of solid Experience Framework in which the special quality of the service experience is considered, (2) need of support for the semantic cohesion between different kinds of data, (3) need of support for the management and search of vast data, and (4) need of building the knowledge base system for collaborative research. Results: HSEAS combines the short information in the customers' words and behaviors or situations and circumstances and provides a place of analysis where the context of the general experience can be read and allows concrete understanding of the actual state and factor of the problem as a Combined Data Analysis Tool. Conclusion: HSEAS becomes the center of information management, analysis and connection and it provides a free collaboration place where physical condition has no relations to as a knowledge base system based on network. Application: It is expected that length and width will be added to the analysis and assistance for effectively accumulating information will be provided in the area of diverse service.

A Keyword Search Model based on the Collected Information of Web Users (웹 사용자 누적 사용정보 기반의 키워드 검색 모델)

  • Yoon, Sung-Hee
    • The Journal of the Korea institute of electronic communication sciences
    • /
    • v.7 no.4
    • /
    • pp.777-782
    • /
    • 2012
  • This paper proposes a technique for improving performance using word senses and user feedback in web information retrieval, compared with the retrieval based on ambiguous user query and index. Disambiguation using query word senses can eliminating the irrelevant pages from the search result. According to semantic categories of nouns which are used as index for retrieval, we build the word sense knowledge-base and categorize the web pages. It can improve the precision of retrieval system with user feedback deciding the query sense and information seeking behavior to pages.

A Study on the Practical Use and Service Implementation of the OPAC 2.0 Based Open API (Open API 기반 OPAC 2.0 서비스 구현 및 유용성에 관한 연구)

  • Gu Jung-Eok;Lee Eung-Bong
    • Journal of the Korean Society for Library and Information Science
    • /
    • v.40 no.2
    • /
    • pp.315-332
    • /
    • 2006
  • In the 21st knowledge society, libraries should consider OPAC redesign and functional performance improvement by using new web related technology. This paper studied and analyzed case studies of Web 2.0. Library 2.0, domestic and international web portals and libraries. Based on the research, this study implemented Open API and RSS reader and proposes its utilization plan as a next generation OPAC 2.0 service model which can improve OPAC search results and search service.

Design and implementation of a EER-based Visual Product Information Modeler (EER기반의 시각적 상품정보 모델링 에디터의 설계와 구현)

  • Tark, Moon-Hee;Kim, Kyung-Hwa;Shim, Jun-Ho
    • The Journal of Society for e-Business Studies
    • /
    • v.12 no.3
    • /
    • pp.97-106
    • /
    • 2007
  • A core technology that may realize the Semantic Web is Ontology. The OWL (Web Ontology Language) has been positioned as a standard language. It requires technical expertise to directly represent the domain knowledge in OWL. Based on our experience of analyzing the fundamental relationships of concepts in e-catalog domain, we have developed a visual product information modeler called PROMOD. The modeling editor makes it possible to automatically generate the OWL codes for the given product information. We employ an Extended Entity-Relationship for conceptual modeling, enriched with modeling elements specialized for the product domain. In this paper, we present our translation schemes from EER model to OWL codes, and how to design and implement the modeling editor. We also provide a scenario to demonstrate the usage of the editor in practice.

  • PDF

An efficient Decision-Making using the extended Fuzzy AHP Method(EFAM) (확장된 Fuzzy AHP를 이용한 효율적인 의사결정)

  • Ryu, Kyung-Hyun;Pi, Su-Young
    • Journal of the Korean Institute of Intelligent Systems
    • /
    • v.19 no.6
    • /
    • pp.828-833
    • /
    • 2009
  • WWW which is an applicable massive set of document on the Web is a thesaurus of various information for users. However, Search engines spend a lot of time to retrieve necessary information and to filter out unnecessary information for user. In this paper, we propose the EFAM(the Extended Fuzzy AHP Method) model to manage the Web resource efficiently, and to make a decision in the problem of specific domain definitely. The EFAM model is concerned with the emotion analysis based on the domain corpus information, and it composed with systematic common concept grids by the knowledge of multiple experts. Therefore, The proposed the EFAM model can extract the documents by considering on the emotion criteria in the semantic context that is extracted concept from the corpus of specific domain and confirms that our model provides more efficient decision-making through an experiment than the conventional methods such as AHP and Fuzzy AHP which describe as a hierarchical structure elements about decision-making based on the alternatives, evaluation criteria, subjective attribute weight and fuzzy relation between concept and object.

Index Ontology Repository for Video Contents (비디오 콘텐츠를 위한 색인 온톨로지 저장소)

  • Hwang, Woo-Yeon;Yang, Jung-Jin
    • Journal of Korea Multimedia Society
    • /
    • v.12 no.10
    • /
    • pp.1499-1507
    • /
    • 2009
  • With the abundance of digital contents, the necessity of precise indexing technology is consistently required. To meet these requirements, the intelligent software entity needs to be the subject of information retrieval and the interoperability among intelligent entities including human must be supported. In this paper, we analyze the unifying framework for multi-modality indexing that Snoek and Worring proposed. Our work investigates the method of improving the authenticity of indexing information in contents-based automated indexing techniques. It supports the creation and control of abstracted high-level indexing information through ontological concepts of Semantic Web skills. Moreover, it attempts to present the fundamental model that allows interoperability between human and machine and between machine and machine. The memory-residence model of processing ontology is inappropriate in order to take-in an enormous amount of indexing information. The use of ontology repository and inference engine is required for consistent retrieval and reasoning of logically expressed knowledge. Our work presents an experiment for storing and retrieving the designed knowledge by using the Minerva ontology repository, which demonstrates satisfied techniques and efficient requirements. At last, the efficient indexing possibility with related research is also considered.

  • PDF

A study on integration of semantic topic based Knowledge model (의미적 토픽 기반 지식모델의 통합에 관한 연구)

  • Chun, Seung-Su;Lee, Sang-Jin;Bae, Sang-Tea
    • Proceedings of the Korean Information Science Society Conference
    • /
    • 2012.06b
    • /
    • pp.181-183
    • /
    • 2012
  • 최근 자연어 및 정형언어 처리, 인공지능 알고리즘 등을 활용한 효율적인 의미 기반 지식모델의 생성과 분석 방법이 제시되고 있다. 이러한 의미 기반 지식모델은 효율적 의사결정트리(Decision Making Tree)와 특정 상황에 대한 체계적인 문제해결(Problem Solving) 경로 분석에 활용된다. 특히 다양한 복잡계 및 사회 연계망 분석에 있어 정적 지표 생성과 회귀 분석, 행위적 모델을 통한 추이분석, 거시예측을 지원하는 모의실험(Simulation) 모형의 기반이 된다. 본 연구에서는 이러한 의미 기반 지식모델을 통합에 있어 텍스트 마이닝을 통해 도출된 토픽(Topic) 모델 간 통합 방법과 정형적 알고리즘을 제시한다. 이를 위해 먼저, 텍스트 마이닝을 통해 도출되는 키워드 맵을 동치적 지식맵으로 변환하고 이를 의미적 지식모델로 통합하는 방법을 설명한다. 또한 키워드 맵으로부터 유의미한 토픽 맵을 투영하는 방법과 의미적 동치 모델을 유도하는 알고리즘을 제안한다. 통합된 의미 기반 지식모델은 토픽 간의 구조적 규칙과 정도 중심성, 근접 중심성, 매개 중심성 등 관계적 의미분석이 가능하며 대규모 비정형 문서의 의미 분석과 활용에 실질적인 기반 연구가 될 수 있다.