• Title/Summary/Keyword: semantic link

Search Result 70, Processing Time 0.031 seconds

Semantic Dependency Link Topic Model for Biomedical Acronym Disambiguation (의미적 의존 링크 토픽 모델을 이용한 생물학 약어 중의성 해소)

  • Kim, Seonho;Yoon, Juntae;Seo, Jungyun
    • Journal of KIISE
    • /
    • v.41 no.9
    • /
    • pp.652-665
    • /
    • 2014
  • Many important terminologies in biomedical text are expressed as abbreviations or acronyms. We newly suggest a semantic link topic model based on the concepts of topic and dependency link to disambiguate biomedical abbreviations and cluster long form variants of abbreviations which refer to the same senses. This model is a generative model inspired by the latent Dirichlet allocation (LDA) topic model, in which each document is viewed as a mixture of topics, with each topic characterized by a distribution over words. Thus, words of a document are generated from a hidden topic structure of a document and the topic structure is inferred from observable word sequences of document collections. In this study, we allow two distinct word generation to incorporate semantic dependencies between words, particularly between expansions (long forms) of abbreviations and their sentential co-occurring words. Besides topic information, the semantic dependency between words is defined as a link and a new random parameter for the link presence is assigned to each word. As a result, the most probable expansions with respect to abbreviations of a given abstract are decided by word-topic distribution, document-topic distribution, and word-link distribution estimated from document collection though the semantic dependency link topic model. The abstracts retrieved from the MEDLINE Entrez interface by the query relating 22 abbreviations and their 186 expansions were used as a data set. The link topic model correctly predicted expansions of abbreviations with the accuracy of 98.30%.

Road Extraction from Images Using Semantic Segmentation Algorithm (영상 기반 Semantic Segmentation 알고리즘을 이용한 도로 추출)

  • Oh, Haeng Yeol;Jeon, Seung Bae;Kim, Geon;Jeong, Myeong-Hun
    • Journal of the Korean Society of Surveying, Geodesy, Photogrammetry and Cartography
    • /
    • v.40 no.3
    • /
    • pp.239-247
    • /
    • 2022
  • Cities are becoming more complex due to rapid industrialization and population growth in modern times. In particular, urban areas are rapidly changing due to housing site development, reconstruction, and demolition. Thus accurate road information is necessary for various purposes, such as High Definition Map for autonomous car driving. In the case of the Republic of Korea, accurate spatial information can be generated by making a map through the existing map production process. However, targeting a large area is limited due to time and money. Road, one of the map elements, is a hub and essential means of transportation that provides many different resources for human civilization. Therefore, it is essential to update road information accurately and quickly. This study uses Semantic Segmentation algorithms Such as LinkNet, D-LinkNet, and NL-LinkNet to extract roads from drone images and then apply hyperparameter optimization to models with the highest performance. As a result, the LinkNet model using pre-trained ResNet-34 as the encoder achieved 85.125 mIoU. Subsequent studies should focus on comparing the results of this study with those of studies using state-of-the-art object detection algorithms or semi-supervised learning-based Semantic Segmentation techniques. The results of this study can be applied to improve the speed of the existing map update process.

DP-LinkNet: A convolutional network for historical document image binarization

  • Xiong, Wei;Jia, Xiuhong;Yang, Dichun;Ai, Meihui;Li, Lirong;Wang, Song
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.15 no.5
    • /
    • pp.1778-1797
    • /
    • 2021
  • Document image binarization is an important pre-processing step in document analysis and archiving. The state-of-the-art models for document image binarization are variants of encoder-decoder architectures, such as FCN (fully convolutional network) and U-Net. Despite their success, they still suffer from three limitations: (1) reduced feature map resolution due to consecutive strided pooling or convolutions, (2) multiple scales of target objects, and (3) reduced localization accuracy due to the built-in invariance of deep convolutional neural networks (DCNNs). To overcome these three challenges, we propose an improved semantic segmentation model, referred to as DP-LinkNet, which adopts the D-LinkNet architecture as its backbone, with the proposed hybrid dilated convolution (HDC) and spatial pyramid pooling (SPP) modules between the encoder and the decoder. Extensive experiments are conducted on recent document image binarization competition (DIBCO) and handwritten document image binarization competition (H-DIBCO) benchmark datasets. Results show that our proposed DP-LinkNet outperforms other state-of-the-art techniques by a large margin. Our implementation and the pre-trained models are available at https://github.com/beargolden/DP-LinkNet.

A Web Link Architecture Based on XRI Providing Persistent Link (영속적 링크를 제공하는 XRI 기반의 웹 링크 구조)

  • Jung, Eui-Hyun;Kim, Weon;Park, Chan-Ki
    • Journal of the Korea Society of Computer and Information
    • /
    • v.13 no.5
    • /
    • pp.247-253
    • /
    • 2008
  • Web 2.0 and Semantic Web technology will be merged to be a next generation Web that leads presentation-oriented Web to data-centric Web. In the next generation Web. semantic processing. Web Platform, and data fusion are most important technology factors. Resolving the Link Rot is the one of the essential technologies to enable these features. The Link Rot causes not only simple annoyances to users but also more serious problems including data integrity. loss of knowledge. breach of service. and so forth. We have suggested a new XRI-based persistent Web link architecture to cure the Link Rot that has been considered as a deep-seated Problem of the Web. The Proposed architecture is based on the XRI suggested by OASIS and it is designed to support a persistent link by using URL rewriting. Since the architecture is designed as a server-side technology, it is superior to existing research especially in Interoperability. Transparency and Adoptability. In addition to this, the architecture provides a metadata identification to be used fer context-aware link resolution.

  • PDF

Research of Adaptive Transformation Method Based on Webpage Semantic Features for Small-Screen Terminals

  • Li, Hao;Liu, Qingtang;Hu, Min;Zhu, Xiaoliang
    • ETRI Journal
    • /
    • v.35 no.5
    • /
    • pp.900-910
    • /
    • 2013
  • Small-screen mobile terminals have difficulty accessing existing Web resources designed for large-screen devices. This paper presents an adaptive transformation method based on webpage semantic features to solve this problem. According to the text density and link density features of the webpages, the webpages are divided into two types: index and content. Our method uses an index-based webpage transformation algorithm and a content-based webpage transformation algorithm. Experiment results demonstrate that our adaptive transformation method is not dependent on specific software and webpage templates, and it is capable of enhancing Web content adaptation on small-screen terminals.

The Design and Implementation of Korean History Web Courseware Using Semantic Network (의미망을 활용한 국사과 웹 코스웨어의 설계 및 구현)

  • Park, Chan-Ghu;Yun, Hong-Won
    • The Journal of Korean Association of Computer Education
    • /
    • v.3 no.1
    • /
    • pp.177-189
    • /
    • 2000
  • This paper describes the design and implementation of Korean History Web courseware using semantic network in order to build learning environment in the viewpoint of cognitive flexibility theory. The most important thing in design for a courseware using semantic network is to build learning environment. The first step to do this is to analyze learning contents and after that we should define the type of link between learning subjects. We should develope the knowledge map which has the link of each type connected with every learning subject.

  • PDF

Content Popularity-Based Peer-to-Peer Semantic Overlay (Content Popularity를 이용한 P2P Semantic Overlay 기법)

  • Choi, Seungbae;Hwang, Euiyoung;Lee, Choonhwa
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2009.11a
    • /
    • pp.523-524
    • /
    • 2009
  • Peer-to-Peer(P2P) 시스템은 분산된 대용량 데이터를 효율적으로 공유하게 하여 사용자들에게 제공되는 killer application 으로 최근까지 여러 분야에서 연구가 되고 있다. 하지만 P2P 네트워크에서 피어가 소유한 데이터나 공통 관심사 또는 사회적인 관계를 고려하지 않고 무작위로 오버레이가 구성되기 때문에 검색 결과의 제약이 발생한다. 따라서 본 논문에서는 P2P 오버레이상의 효율적인 데이터 검색을 위해서 각 피어가 가지고 있는 데이터와 공통의 관심사를 기반으로 유사성을 측정하여 semantic overlay를 구성하는 기법을 제안한다. 그리고 피어들 간의 semantic proximity는 데이터 요약 기법을 사용하여 측정되며 측정 과정상에서 popular content을 고려하여 semantic proximity의 왜곡현상을 방지하여 semantic link quality의 향상을 가져오는 방안을 도입한다.

A Ranking Algorithm for Semantic Web Resources: A Class-oriented Approach (시맨틱 웹 자원의 랭킹을 위한 알고리즘: 클래스중심 접근방법)

  • Rho, Sang-Kyu;Park, Hyun-Jung;Park, Jin-Soo
    • Asia pacific journal of information systems
    • /
    • v.17 no.4
    • /
    • pp.31-59
    • /
    • 2007
  • We frequently use search engines to find relevant information in the Web but still end up with too much information. In order to solve this problem of information overload, ranking algorithms have been applied to various domains. As more information will be available in the future, effectively and efficiently ranking search results will become more critical. In this paper, we propose a ranking algorithm for the Semantic Web resources, specifically RDF resources. Traditionally, the importance of a particular Web page is estimated based on the number of key words found in the page, which is subject to manipulation. In contrast, link analysis methods such as Google's PageRank capitalize on the information which is inherent in the link structure of the Web graph. PageRank considers a certain page highly important if it is referred to by many other pages. The degree of the importance also increases if the importance of the referring pages is high. Kleinberg's algorithm is another link-structure based ranking algorithm for Web pages. Unlike PageRank, Kleinberg's algorithm utilizes two kinds of scores: the authority score and the hub score. If a page has a high authority score, it is an authority on a given topic and many pages refer to it. A page with a high hub score links to many authoritative pages. As mentioned above, the link-structure based ranking method has been playing an essential role in World Wide Web(WWW), and nowadays, many people recognize the effectiveness and efficiency of it. On the other hand, as Resource Description Framework(RDF) data model forms the foundation of the Semantic Web, any information in the Semantic Web can be expressed with RDF graph, making the ranking algorithm for RDF knowledge bases greatly important. The RDF graph consists of nodes and directional links similar to the Web graph. As a result, the link-structure based ranking method seems to be highly applicable to ranking the Semantic Web resources. However, the information space of the Semantic Web is more complex than that of WWW. For instance, WWW can be considered as one huge class, i.e., a collection of Web pages, which has only a recursive property, i.e., a 'refers to' property corresponding to the hyperlinks. However, the Semantic Web encompasses various kinds of classes and properties, and consequently, ranking methods used in WWW should be modified to reflect the complexity of the information space in the Semantic Web. Previous research addressed the ranking problem of query results retrieved from RDF knowledge bases. Mukherjea and Bamba modified Kleinberg's algorithm in order to apply their algorithm to rank the Semantic Web resources. They defined the objectivity score and the subjectivity score of a resource, which correspond to the authority score and the hub score of Kleinberg's, respectively. They concentrated on the diversity of properties and introduced property weights to control the influence of a resource on another resource depending on the characteristic of the property linking the two resources. A node with a high objectivity score becomes the object of many RDF triples, and a node with a high subjectivity score becomes the subject of many RDF triples. They developed several kinds of Semantic Web systems in order to validate their technique and showed some experimental results verifying the applicability of their method to the Semantic Web. Despite their efforts, however, there remained some limitations which they reported in their paper. First, their algorithm is useful only when a Semantic Web system represents most of the knowledge pertaining to a certain domain. In other words, the ratio of links to nodes should be high, or overall resources should be described in detail, to a certain degree for their algorithm to properly work. Second, a Tightly-Knit Community(TKC) effect, the phenomenon that pages which are less important but yet densely connected have higher scores than the ones that are more important but sparsely connected, remains as problematic. Third, a resource may have a high score, not because it is actually important, but simply because it is very common and as a consequence it has many links pointing to it. In this paper, we examine such ranking problems from a novel perspective and propose a new algorithm which can solve the problems under the previous studies. Our proposed method is based on a class-oriented approach. In contrast to the predicate-oriented approach entertained by the previous research, a user, under our approach, determines the weights of a property by comparing its relative significance to the other properties when evaluating the importance of resources in a specific class. This approach stems from the idea that most queries are supposed to find resources belonging to the same class in the Semantic Web, which consists of many heterogeneous classes in RDF Schema. This approach closely reflects the way that people, in the real world, evaluate something, and will turn out to be superior to the predicate-oriented approach for the Semantic Web. Our proposed algorithm can resolve the TKC(Tightly Knit Community) effect, and further can shed lights on other limitations posed by the previous research. In addition, we propose two ways to incorporate data-type properties which have not been employed even in the case when they have some significance on the resource importance. We designed an experiment to show the effectiveness of our proposed algorithm and the validity of ranking results, which was not tried ever in previous research. We also conducted a comprehensive mathematical analysis, which was overlooked in previous research. The mathematical analysis enabled us to simplify the calculation procedure. Finally, we summarize our experimental results and discuss further research issues.

An Improved Position Estimation Algorithm of Vehicles Using Semantic Information of Maps (지도의 의미 정보를 이용한 개선된 차량 위치 추정 알고리즘)

  • Lee, Chang Gil;Choi, Yoon Ho;Park, Jin Bae
    • Journal of Institute of Control, Robotics and Systems
    • /
    • v.22 no.9
    • /
    • pp.753-758
    • /
    • 2016
  • In this paper, we propose a novel method for estimating a vehicle's current position, even on roads that have similar patterns. In the proposed method, we classified the semantic information of the nodes in detail and added the semantic information of the link to solve the problem due to similar and repeated patterns. We also improved the mapping method by comparing the result of the duplicated matching with that of the only matching obtained just before corresponding duplicated matching. From the simulation results, we verify that the performance of the proposed method is better than that of the existing method.

Linked Legal Data Construction and Connection of LOD Cloud

  • Jo, Dae Woong;Kim, Myung Ho
    • Journal of the Korea Society of Computer and Information
    • /
    • v.21 no.5
    • /
    • pp.11-18
    • /
    • 2016
  • Linked Data is a web standard data definition method devised to connect, expand resources with a standardized type. Linked Data built in various areas expands existing knowledge through an open data cloud like LOD(Linked Open Data). A project to link and service existing knowledge through LOD is under way worldwide. However, LOD project in domestic is being participated in a specific field to the level of research. In this paper, we suggests a method to build the area of technical knowledge like legislations in type of Linked Data, and distribute such Linked Data built to LOD. The construction method suggested by this paper divides knowledge of legislations in structural, semantic, and integrated perspective, and builds each of them by converting to Linked Data according to the perspective. Also, such built Linked Legal Data prepares to link knowledge in a standardized type by distributing them onto LOD. Built Linked Legal Data are equipped with schema for link service in various types, and give help increase understand the access type to existing legal information.