• Title/Summary/Keyword: 키워드 추출 방법

Search Result 355, Processing Time 0.031 seconds

A Systematic Review on Measurement Instruments of Bilateral Upper Extremity Function (양측 상지 기능 평가도구에 관한 체계적 고찰)

  • Lee, Joo-Hyun;Lee, Ye-Jin;Park, Ji-Hyuk
    • Therapeutic Science for Rehabilitation
    • /
    • v.5 no.1
    • /
    • pp.7-22
    • /
    • 2016
  • Objective : This study was conducted to review about instrument for bilateral upper extremity assessment. Methods : We searched published papers in Medline database. The keywords used in the search were 'upper extremity' and 'motor activity', 'activities of daily living' and 'assessment', 'instrument', 'disability evaluations'. In total of 68 papers, 44 assessment instruments was extracted. We analyzed about numbers, subjects, methods, reliability, validity, responsiveness of assessment instruments about bilateral upper limb function comparing unilateral, combined bilateral and unilateral instruments. Results : The numbers of bilateral measurement tool were 2 in a total of upper extremity measurement tools. Also, subjects were patients with stroke and measurement was based on performance. The reliability, validity and responsiveness of tools were high. Conclusions : Futher study will be needed to development and research about instrument of bilateral upper extremity.

Medical Image Automatic Annotation Using Multi-class SVM and Annotation Code Array (다중 클래스 SVM과 주석 코드 배열을 이용한 의료 영상 자동 주석 생성)

  • Park, Ki-Hee;Ko, Byoung-Chul;Nam, Jae-Yeal
    • The KIPS Transactions:PartB
    • /
    • v.16B no.4
    • /
    • pp.281-288
    • /
    • 2009
  • This paper proposes a novel algorithm for the efficient classification and annotation of medical images, especially X-ray images. Since X-ray images have a bright foreground against a dark background, we need to extract the different visual descriptors compare with general nature images. In this paper, a Color Structure Descriptor (CSD) based on Harris Corner Detector is only extracted from salient points, and an Edge Histogram Descriptor (EHD) used for a textual feature of image. These two feature vectors are then applied to a multi-class Support Vector Machine (SVM), respectively, to classify images into one of 20 categories. Finally, an image has the Annotation Code Array based on the pre-defined hierarchical relations of categories and priority code order, which is given the several optimal keywords by the Annotation Code Array. Our experiments show that our annotation results have better annotation performance when compared to other method.

The Design of Blog Network Analysis System using Map/Reduce Programming Model (Map/Reduce를 이용한 블로그 연결망 분석 시스템 설계)

  • Joe, In-Whee;Park, Jae-Kyun
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.35 no.9B
    • /
    • pp.1259-1265
    • /
    • 2010
  • Recently, on-line social network has been increasing according to development of internet. The most representative service is blog. A Blog is a type of personal web site, usually maintained by an individual with regular entries of commentary. These blogs are related to each other, and it is called Blog Network in this paper. In a blog network, posts in a blog can be diffused to other blogs. Analyzing information diffusion in a blog world is a very useful research issue, which can be used for predicting information diffusion, abnormally detection, marketing, and revitalizing the blog world. Existing studies on network analysis have no consideration for the passage of time and these approaches can only measure network activity for a node by the number of direct connections that a given node has. As one solution, this paper suggests the new method of measuring the blog network activity using logistic curve model and Cosine-similarity in key words by the Map/Reduce programming model.

A Study on Research Paper Classification Using Keyword Clustering (키워드 군집화를 이용한 연구 논문 분류에 관한 연구)

  • Lee, Yun-Soo;Pheaktra, They;Lee, JongHyuk;Gil, Joon-Min
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.7 no.12
    • /
    • pp.477-484
    • /
    • 2018
  • Due to the advancement of computer and information technologies, numerous papers have been published. As new research fields continue to be created, users have a lot of trouble finding and categorizing their interesting papers. In order to alleviate users' this difficulty, this paper presents a method of grouping similar papers and clustering them. The presented method extracts primary keywords from the abstracts of each paper by using TF-IDF. Based on TF-IDF values extracted using K-means clustering algorithm, our method clusters papers to the ones that have similar contents. To demonstrate the practicality of the proposed method, we use paper data in FGCS journal as actual data. Based on these data, we derive the number of clusters using Elbow scheme and show clustering performance using Silhouette scheme.

Research on analysis of articleable advertisements and design of extraction method for articleable advertisements using deep learning

  • Seoksoo Kim;Jae-Young Jung
    • Journal of the Korea Society of Computer and Information
    • /
    • v.29 no.6
    • /
    • pp.13-22
    • /
    • 2024
  • There is a need for and positive aspects of article-based advertising, but as exaggerated and disguised information is delivered due to some indiscriminate 'article-based advertisements', readers have difficulty distinguishing between general articles and article-based advertisements, leading to a lot of misinterpretation and confusion of information. is doing Since readers will continue to acquire new information and apply this information at the right time and place to bring a lot of value, it is judged to be even more important to distinguish between accurate general articles and article-like advertisements. Therefore, as differentiated information between general articles and article-like advertisements is needed, as part of this, for readers who have difficulty identifying accurate information due to such indiscriminate article-like advertisements in Internet newspapers, this paper introduces IT and AI technologies. We attempted to present a method that can be solved in terms of a system that incorporates, and this method was designed to extract articleable advertisements using a knowledge-based natural language processing method that finds and refines advertising keywords and deep learning technology.

A study on the classification of research topics based on COVID-19 academic research using Topic modeling (토픽모델링을 활용한 COVID-19 학술 연구 기반 연구 주제 분류에 관한 연구)

  • Yoo, So-yeon;Lim, Gyoo-gun
    • Journal of Intelligence and Information Systems
    • /
    • v.28 no.1
    • /
    • pp.155-174
    • /
    • 2022
  • From January 2020 to October 2021, more than 500,000 academic studies related to COVID-19 (Coronavirus-2, a fatal respiratory syndrome) have been published. The rapid increase in the number of papers related to COVID-19 is putting time and technical constraints on healthcare professionals and policy makers to quickly find important research. Therefore, in this study, we propose a method of extracting useful information from text data of extensive literature using LDA and Word2vec algorithm. Papers related to keywords to be searched were extracted from papers related to COVID-19, and detailed topics were identified. The data used the CORD-19 data set on Kaggle, a free academic resource prepared by major research groups and the White House to respond to the COVID-19 pandemic, updated weekly. The research methods are divided into two main categories. First, 41,062 articles were collected through data filtering and pre-processing of the abstracts of 47,110 academic papers including full text. For this purpose, the number of publications related to COVID-19 by year was analyzed through exploratory data analysis using a Python program, and the top 10 journals under active research were identified. LDA and Word2vec algorithm were used to derive research topics related to COVID-19, and after analyzing related words, similarity was measured. Second, papers containing 'vaccine' and 'treatment' were extracted from among the topics derived from all papers, and a total of 4,555 papers related to 'vaccine' and 5,971 papers related to 'treatment' were extracted. did For each collected paper, detailed topics were analyzed using LDA and Word2vec algorithms, and a clustering method through PCA dimension reduction was applied to visualize groups of papers with similar themes using the t-SNE algorithm. A noteworthy point from the results of this study is that the topics that were not derived from the topics derived for all papers being researched in relation to COVID-19 (

    ) were the topic modeling results for each research topic (
    ) was found to be derived from For example, as a result of topic modeling for papers related to 'vaccine', a new topic titled Topic 05 'neutralizing antibodies' was extracted. A neutralizing antibody is an antibody that protects cells from infection when a virus enters the body, and is said to play an important role in the production of therapeutic agents and vaccine development. In addition, as a result of extracting topics from papers related to 'treatment', a new topic called Topic 05 'cytokine' was discovered. A cytokine storm is when the immune cells of our body do not defend against attacks, but attack normal cells. Hidden topics that could not be found for the entire thesis were classified according to keywords, and topic modeling was performed to find detailed topics. In this study, we proposed a method of extracting topics from a large amount of literature using the LDA algorithm and extracting similar words using the Skip-gram method that predicts the similar words as the central word among the Word2vec models. The combination of the LDA model and the Word2vec model tried to show better performance by identifying the relationship between the document and the LDA subject and the relationship between the Word2vec document. In addition, as a clustering method through PCA dimension reduction, a method for intuitively classifying documents by using the t-SNE technique to classify documents with similar themes and forming groups into a structured organization of documents was presented. In a situation where the efforts of many researchers to overcome COVID-19 cannot keep up with the rapid publication of academic papers related to COVID-19, it will reduce the precious time and effort of healthcare professionals and policy makers, and rapidly gain new insights. We hope to help you get It is also expected to be used as basic data for researchers to explore new research directions.

  • Methods for Integration of Documents using Hierarchical Structure Representation based on the Formal Concept Analysis (FCA 기반 계층적 구조 표현을 이용한 문서 통합 기법)

    • Kim, Tae-Hwan;Park, Jae-Hyun;Choi, Joong-Min
      • Proceedings of the Korean Information Science Society Conference
      • /
      • 2006.10b
      • /
      • pp.388-392
      • /
      • 2006
    • 가공해서 사용하는 정보량이 많아질수록 원하는 정보를 찾는 데 더 많은 노력이 필요하게 마련이다. 따라서 사람들은 대대로 정보를 구조화하는 방법들을 고안해왔으며, 여러 가지 계층적 구조화 방법들을 사용했었다. 이렇게 구현된 정보의 계층 구조는 키워드 검색을 바탕으로 수평적 계층 구조만을 가지는 구조였다. 자료가 전문화되고 정보를 검색하는 사용자 또한 검색된 정보와 관련된 정보를 더 원하는 현 시점에서 정보의 수평적 계층 구조만으로 사용자의 만족도를 충족할 수 없다. 이러한 문제점을 해결하기 위해 이 논문에서는 특정 도메인의 문서를 단락별 명사와 동사 및 목적어를 추출하여 해당 동사가 명사 및 목적어를 취할 수 있는 가능한 값을 체크하여 그 단락의 계층적 트리를 구성하고, 단락별 트리를 이용하여 문서의 내용을 트리로 재구성할 수 있게 된다. 이렇게 만들어진 문서의 트리들은 트리의 구조를 보고 특정 문서에 더 구체적인지 아니면 더 일반적인지 측정하여 문서와 문서간의 관계 또한 트리 형식으로 보여주어 사용자가 원하는 정보를 보다 쉽게 검색해 주는 자동화 문서 계층 구조를 제안한다.

    • PDF

    Topic-Network based Topic Shift Detection on Twitter (트위터 데이터를 이용한 네트워크 기반 토픽 변화 추적 연구)

    • Jin, Seol A;Heo, Go Eun;Jeong, Yoo Kyung;Song, Min
      • Journal of the Korean Society for information Management
      • /
      • v.30 no.1
      • /
      • pp.285-302
      • /
      • 2013
    • This study identified topic shifts and patterns over time by analyzing an enormous amount of Twitter data whose characteristics are high accessibility and briefness. First, we extracted keywords for a certain product and used them for representing the topic network allows for intuitive understanding of keywords associated with topics by nodes and edges by co-word analysis. We conducted temporal analysis of term co-occurrence as well as topic modeling to examine the results of network analysis. In addition, the results of comparing topic shifts on Twitter with the corresponding retrieval results from newspapers confirm that Twitter makes immediate responses to news media and spreads the negative issues out quickly. Our findings may suggest that companies utilize the proposed technique to identify public's negative opinions as quickly as possible and to apply for the timely decision making and effective responses to their customers.

    Research Outcomes and Limitations of Records and Archives Organization in Korea (국내 기록조직 연구의 성과와 과제)

    • Lee, Eun-Ju;Rho, Jee-Hyun
      • Journal of Korean Society of Archives and Records Management
      • /
      • v.20 no.4
      • /
      • pp.129-146
      • /
      • 2020
    • This study aims to investigate the outcomes and limitations of research studies on records and archives organization published in Korea. In particular, it will serve as an in-depth examination of the contribution of this area of research to the improvements and changes in the country's records management field. To this end, 150 journal articles related to the records and archives organization were gathered. After extracting refined keywords from the titles and author-assigned keywords, terminology analysis and contents analysis were conducted. On the one hand, terminology analysis (frequency and network analysis) identified frequently discussed topics and the relationships between them. On the other hand, through content analysis, the study revealed the detailed contents regarding the two main topics and their meanings.

    An Analysis of Keywords Related to Neighborhood Healing Gardens Using Big Data (빅데이터를 활용한 생활밀착형 치유정원 연관키워드 분석)

    • Huang, Zhirui;Lee, Ai-Ran
      • Land and Housing Review
      • /
      • v.13 no.2
      • /
      • pp.81-90
      • /
      • 2022
    • This study is based on social needs for green healing spaces assumed to enhance mental health in a city. This study proposes development directions through the analysis of modern social recognition factors for neighborhood gardens. As a research method, web information data was collected using Textom among big data tools. Text Mining was conducted to extract elements and analyze their relationship through keyword analysis, network analysis, and cluster analysis. As a result, first, the healing space and the healing environment were creating an eco-friendly healthy environment in a space close to the neighborhood within the city. Second, neighborhood gardens included projects and activities that involved government, local administration, and citizens by linking facilities as well as living culture and urban environments. These gardens have been reinforced through green welfare and service programs. In conclusion, friendly gardens in the neighborhood for the purpose of public interest, which are beneficial to mental health, are green infrastructures as a healing environment that can produce positive effects.


    (34141) Korea Institute of Science and Technology Information, 245, Daehak-ro, Yuseong-gu, Daejeon
    Copyright (C) KISTI. All Rights Reserved.