• Title/Summary/Keyword: RDF(S)

Search Result 128, Processing Time 0.025 seconds

Semantic-based Keyword Search System over Relational Database (관계형 데이터베이스에서의 시맨틱 기반 키워드 탐색 시스템)

  • Yang, Younghyoo
    • Journal of the Korea Society of Computer and Information
    • /
    • v.18 no.12
    • /
    • pp.91-101
    • /
    • 2013
  • One issue with keyword search in general is its ambiguity which can ultimately impact the effectiveness of the search in terms of the quality of the search results. This ambiguity is primarily due to the ambiguity of the contextual meaning of each term in the query. In addition to the query ambiguity itself, the relationships between the keywords in the search results are crucial for the proper interpretation of the search results by the user and should be clearly presented in the search results. We address the keyword search ambiguity issue by adapting some of the existing approaches for keyword mapping from the query terms to the schema terms/instances. The approaches we have adapted for term mapping capture both the syntactic similarity between the query keywords and the schema terms as well as the semantic similarity of the two and give better mappings and ultimately 50% raised accurate results. Finally, to address the last issue of lacking clear relationships among the terms appearing in the search results, our system has leveraged semantic web technologies in order to enrich the knowledgebase and to discover the relationships between the keywords.

Development of Forward chaining inference engine SMART-F using Rete Algorithm in the Semantic Web (차세대 웹 환경에서의 Rete Algorithm을 이용한 정방향 추론엔진 SMART - F 개발)

  • Jeong, Kyun-Beom;Hong, June-Seok;Kim, Woo-Ju;Lee, Myung-Jin;Park, Ji-Hyoung;Song, Yong-Uk
    • Journal of Intelligence and Information Systems
    • /
    • v.13 no.3
    • /
    • pp.17-29
    • /
    • 2007
  • Inference engine that performs the brain of software agent in next generation's web with various standards based on standard language of the web, XML has to understand SWRL (Semantic Web Rule Language) that is a language to express the rule in the Semantic Web. In this research, we want to develop a forward inference engine, SMART-F (SeMantic web Agent Reasoning Tools-Forward chaining inference engine) that uses SWRL as a rule express method, and OWL as a fact express method. In the traditional inference field, the Rete algorithm that improves effectiveness of forward rule inference by converting if-then rules to network structure is often used for forward inference. To apply this to the Semantic Web, we analyze the required functions for the SWRL-based forward inference, and design the forward inference algorithm that reflects required functions of next generation's Semantic Web deducted by Rete algorithm. And then, to secure each platform's independence and portability in the ubiquitous environment and overcome the gap of performance, we developed management tool of fact and rule base and forward inference engine. This is compatible with fact and rule base of SMART-B that was developed. So, this maximizes a practical use of knowledge in the next generation's Web environment.

  • PDF

Scalable RDFS Reasoning using Logic Programming Approach in a Single Machine (단일머신 환경에서의 논리적 프로그래밍 방식 기반 대용량 RDFS 추론 기법)

  • Jagvaral, Batselem;Kim, Jemin;Lee, Wan-Gon;Park, Young-Tack
    • Journal of KIISE
    • /
    • v.41 no.10
    • /
    • pp.762-773
    • /
    • 2014
  • As the web of data is increasingly producing large RDFS datasets, it becomes essential in building scalable reasoning engines over large triples. There have been many researches used expensive distributed framework, such as Hadoop, to reason over large RDFS triples. However, in many cases we are required to handle millions of triples. In such cases, it is not necessary to deploy expensive distributed systems because logic program based reasoners in a single machine can produce similar reasoning performances with that of distributed reasoner using Hadoop. In this paper, we propose a scalable RDFS reasoner using logical programming methods in a single machine and compare our empirical results with that of distributed systems. We show that our logic programming based reasoner using a single machine performs as similar as expensive distributed reasoner does up to 200 million RDFS triples. In addition, we designed a meta data structure by decomposing the ontology triples into separate sectors. Instead of loading all the triples into a single model, we selected an appropriate subset of the triples for each ontology reasoning rule. Unification makes it easy to handle conjunctive queries for RDFS schema reasoning, therefore, we have designed and implemented RDFS axioms using logic programming unifications and efficient conjunctive query handling mechanisms. The throughputs of our approach reached to 166K Triples/sec over LUBM1500 with 200 million triples. It is comparable to that of WebPIE, distributed reasoner using Hadoop and Map Reduce, which performs 185K Triples/sec. We show that it is unnecessary to use the distributed system up to 200 million triples and the performance of logic programming based reasoner in a single machine becomes comparable with that of expensive distributed reasoner which employs Hadoop framework.

Anaerobic Bacterial Degradation for the Effective Utilization of Biomass

  • Ohmiya, Kunio;Sakka, Kazuo;Kimura, Tetsuya
    • Biotechnology and Bioprocess Engineering:BBE
    • /
    • v.10 no.6
    • /
    • pp.482-493
    • /
    • 2005
  • Biomass is originally photosynthesized from inorgainic compounds such as $CO_2$, minerals, water and solar energy. Recent studies have shown that anaerobic bacteria have the ability to convert recalcitrant biomass such as cellullosic or chitinoic materials to useful compounds. The biomass containing agricultural waste, unutilized wood and other garbage is expected to utilize as feed, food and fuel by microbial degradation and other metabolic functions. In this study we isolated several anaerobic, cellulolytic and chitinolytic bacteria from rumen fluid, compost and soil to study their related enzymes and genes. The anaerobic and cellulolytic bacteria, Clostridium thermocellum, Clostridium stercorarium, and Clostridium josui, were isolated from compost and the chitinolytic Clostridium paraputrificum from beach soil and Ruminococcus albus was isolated from cow rumen. After isolation, novel cellulase and xylanase genes from these anaerobes were cloned and expressed in Escherichia coli. The properties of the cloned enzymes showed that some of them were the components of the enzyme (cellulase) complex, i.e., cellulosome, which is known to form complexes by binding cohesin domains on the cellulase integrating protein (Cip: or core protein) and dockerin domains on the enzymes. Several dockerin and cohesin polypeptides were independently produced by E. coli and their binding properties were specified with BIAcore by measuring surface plasmon resonance. Three pairs of cohesin-dockerin with differing binding specificities were selected. Two of their genes encoding their respective cohesin polypeptides were combined to one gene and expressed in E. coli as a chimeric core protein, on which two dockerin-dehydrogenase chimeras, the dockerin-formaldehyde dehydrogenase and the dockerin-NADH dehydrogenase are planning to bind for catalyzing $CO_2$ reduction to formic acid by feeding NADH. This reaction may represent a novel strategy for the reduction of the green house gases. Enzymes from the anaerobes were also expressed in tobacco and rice plants. The activity of a xylanase from C. stercorarium was detected in leaves, stems, and rice grain under the control of CaMV35S promoter. The digestibility of transgenic rice leaves in goat rumen was slightly accelerated. C. paraputrificum was found to solubilize shrimp shells and chitin to generate hydrogen gas. Hydrogen productivity (1.7 mol $H_2/mol$ glucos) of the organism was improved up to 1.8 times by additional expression of the own hydrogenase gene in C. paraputrficum using a modified vector of Clostridiu, perfringens. The hydrygen producing microflora from soil, garbage and dried pelletted garbage, known as refuse derived fuel(RDF), were also found to be effective in converting biomass waste to hydrogen gas.

Automatic Recommendation of Nearby Tourist Attractions related to Events (이벤트와 관련된 주변 관광지 자동 추천 알고리즘 개발)

  • Ahn, Jinhyun;Im, Dong-Hyuk
    • Journal of the Korea Academia-Industrial cooperation Society
    • /
    • v.21 no.3
    • /
    • pp.407-413
    • /
    • 2020
  • Participating in exhibitions is one of the major activities for tourists. When selecting their next travel destination after participating in an event, they use map services and social network services, such as blogs, to obtain information about tourist attractions. The map services are location-based recommendations, because they can easily retrieve information regarding nearby places. Blogs contain informative content about tourist attractions, thereby providing content-based recommendations. However, few services consider both location and content. In location-based recommendations, tourist attractions that are not related to the content of the event attended might be recommended. Content-based recommendation has a disadvantage in that events located at a distance might get recommended. We propose an algorithm that considers both location and content, based on information from the Korea Tourism Organization's Linked Open Data (LOD), Wikipedia, and a Korean dictionary. By extracting nouns from the description of a tourist attraction and then comparing them with nouns about other attractions, a content-based relationship is determined. The distance to the event is calculated based on the latitude and longitude of each tourist attraction. A weight selected by the user is used for linear combination with the content-based relationship to determine the preference order of the recommendations.

A Dynamic Management Method for FOAF Using RSS and OLAP cube (RSS와 OLAP 큐브를 이용한 FOAF의 동적 관리 기법)

  • Sohn, Jong-Soo;Chung, In-Jeong
    • Journal of Intelligence and Information Systems
    • /
    • v.17 no.2
    • /
    • pp.39-60
    • /
    • 2011
  • Since the introduction of web 2.0 technology, social network service has been recognized as the foundation of an important future information technology. The advent of web 2.0 has led to the change of content creators. In the existing web, content creators are service providers, whereas they have changed into service users in the recent web. Users share experiences with other users improving contents quality, thereby it has increased the importance of social network. As a result, diverse forms of social network service have been emerged from relations and experiences of users. Social network is a network to construct and express social relations among people who share interests and activities. Today's social network service has not merely confined itself to showing user interactions, but it has also developed into a level in which content generation and evaluation are interacting with each other. As the volume of contents generated from social network service and the number of connections between users have drastically increased, the social network extraction method becomes more complicated. Consequently the following problems for the social network extraction arise. First problem lies in insufficiency of representational power of object in the social network. Second problem is incapability of expressional power in the diverse connections among users. Third problem is the difficulty of creating dynamic change in the social network due to change in user interests. And lastly, lack of method capable of integrating and processing data efficiently in the heterogeneous distributed computing environment. The first and last problems can be solved by using FOAF, a tool for describing ontology-based user profiles for construction of social network. However, solving second and third problems require a novel technology to reflect dynamic change of user interests and relations. In this paper, we propose a novel method to overcome the above problems of existing social network extraction method by applying FOAF (a tool for describing user profiles) and RSS (a literary web work publishing mechanism) to OLAP system in order to dynamically innovate and manage FOAF. We employed data interoperability which is an important characteristic of FOAF in this paper. Next we used RSS to reflect such changes as time flow and user interests. RSS, a tool for literary web work, provides standard vocabulary for distribution at web sites and contents in the form of RDF/XML. In this paper, we collect personal information and relations of users by utilizing FOAF. We also collect user contents by utilizing RSS. Finally, collected data is inserted into the database by star schema. The system we proposed in this paper generates OLAP cube using data in the database. 'Dynamic FOAF Management Algorithm' processes generated OLAP cube. Dynamic FOAF Management Algorithm consists of two functions: one is find_id_interest() and the other is find_relation (). Find_id_interest() is used to extract user interests during the input period, and find-relation() extracts users matching user interests. Finally, the proposed system reconstructs FOAF by reflecting extracted relationships and interests of users. For the justification of the suggested idea, we showed the implemented result together with its analysis. We used C# language and MS-SQL database, and input FOAF and RSS as data collected from livejournal.com. The implemented result shows that foaf : interest of users has reached an average of 19 percent increase for four weeks. In proportion to the increased foaf : interest change, the number of foaf : knows of users has grown an average of 9 percent for four weeks. As we use FOAF and RSS as basic data which have a wide support in web 2.0 and social network service, we have a definite advantage in utilizing user data distributed in the diverse web sites and services regardless of language and types of computer. By using suggested method in this paper, we can provide better services coping with the rapid change of user interests with the automatic application of FOAF.

Size Distribution and Physicochemical Characteristics of MSW for Design of Its Mechanical Biological Treatment Process (폐기물전처리(MBT)시설 설계를 위한 생활폐기물의 입도분포 및 물리화학적 특성에 관한 연구)

  • Park, Jin-Kyu;Song, Sang-Hoon;Jeong, Sae-Rom;Jung, Min-Soo;Lee, Nam-Hoon;Lee, Byoung-Chul
    • Journal of the Korea Organic Resources Recycling Association
    • /
    • v.16 no.1
    • /
    • pp.62-69
    • /
    • 2008
  • There has been a recent trend in Korea that treatments for combustible wastes among municipal solid waste (MSW) by those methods, such as incineration and landfill are restricted as much as possible and Mechanical Biological Treatment (MBT) are encouraged actively in order to promote resource recovery. To build and operate properly these facilities, the physicochemical characteristics of MSW should be analyzed precisely beforehand. In particular, designing a crusher or separator properly which is the main process in MBT facilities of MSW. require the information on the size distribution characteristics of MSW, but they are nor sufficient in the qualities and quantities yet as of now. Accordingly, this study aims to evaluate size distribution characteristics of MSW and its physicochemical characteristics by size. The samples of MSW were collected from detached dwelling area, apartment area, business area, and commercial area of A city in Korea. According to the result of analysis, paper records 29.78~60.02% by wet weight basis, so it was the most regardless of the regions where the wastes were generated. And in terms of element analysis, Carbon(C) was 34.77~44.39%, the largest friction, and Oxygen(O) was the next occupying 19.46~33.71%. As indices of RDFs, Chlorine(Cl) was 0.39~0.83%, so it was less than the standard, 2.0%(by dry weight basis); moreover, Sulfur(S) did not exceed the standard, 0.6%, either. In the size distribution of MSW, waste fraction ranging 50~80mm in diameter was the most in combustible waste while 30~50mm was in incombustible waste.

  • PDF

Knowledge Extraction Methodology and Framework from Wikipedia Articles for Construction of Knowledge-Base (지식베이스 구축을 위한 한국어 위키피디아의 학습 기반 지식추출 방법론 및 플랫폼 연구)

  • Kim, JaeHun;Lee, Myungjin
    • Journal of Intelligence and Information Systems
    • /
    • v.25 no.1
    • /
    • pp.43-61
    • /
    • 2019
  • Development of technologies in artificial intelligence has been rapidly increasing with the Fourth Industrial Revolution, and researches related to AI have been actively conducted in a variety of fields such as autonomous vehicles, natural language processing, and robotics. These researches have been focused on solving cognitive problems such as learning and problem solving related to human intelligence from the 1950s. The field of artificial intelligence has achieved more technological advance than ever, due to recent interest in technology and research on various algorithms. The knowledge-based system is a sub-domain of artificial intelligence, and it aims to enable artificial intelligence agents to make decisions by using machine-readable and processible knowledge constructed from complex and informal human knowledge and rules in various fields. A knowledge base is used to optimize information collection, organization, and retrieval, and recently it is used with statistical artificial intelligence such as machine learning. Recently, the purpose of the knowledge base is to express, publish, and share knowledge on the web by describing and connecting web resources such as pages and data. These knowledge bases are used for intelligent processing in various fields of artificial intelligence such as question answering system of the smart speaker. However, building a useful knowledge base is a time-consuming task and still requires a lot of effort of the experts. In recent years, many kinds of research and technologies of knowledge based artificial intelligence use DBpedia that is one of the biggest knowledge base aiming to extract structured content from the various information of Wikipedia. DBpedia contains various information extracted from Wikipedia such as a title, categories, and links, but the most useful knowledge is from infobox of Wikipedia that presents a summary of some unifying aspect created by users. These knowledge are created by the mapping rule between infobox structures and DBpedia ontology schema defined in DBpedia Extraction Framework. In this way, DBpedia can expect high reliability in terms of accuracy of knowledge by using the method of generating knowledge from semi-structured infobox data created by users. However, since only about 50% of all wiki pages contain infobox in Korean Wikipedia, DBpedia has limitations in term of knowledge scalability. This paper proposes a method to extract knowledge from text documents according to the ontology schema using machine learning. In order to demonstrate the appropriateness of this method, we explain a knowledge extraction model according to the DBpedia ontology schema by learning Wikipedia infoboxes. Our knowledge extraction model consists of three steps, document classification as ontology classes, proper sentence classification to extract triples, and value selection and transformation into RDF triple structure. The structure of Wikipedia infobox are defined as infobox templates that provide standardized information across related articles, and DBpedia ontology schema can be mapped these infobox templates. Based on these mapping relations, we classify the input document according to infobox categories which means ontology classes. After determining the classification of the input document, we classify the appropriate sentence according to attributes belonging to the classification. Finally, we extract knowledge from sentences that are classified as appropriate, and we convert knowledge into a form of triples. In order to train models, we generated training data set from Wikipedia dump using a method to add BIO tags to sentences, so we trained about 200 classes and about 2,500 relations for extracting knowledge. Furthermore, we evaluated comparative experiments of CRF and Bi-LSTM-CRF for the knowledge extraction process. Through this proposed process, it is possible to utilize structured knowledge by extracting knowledge according to the ontology schema from text documents. In addition, this methodology can significantly reduce the effort of the experts to construct instances according to the ontology schema.