• Title/Summary/Keyword: Korean parsing

Search Result 326, Processing Time 0.028 seconds

A Study on Smartwatch review data of SNS and sentiment analytical using opinion mining (스마트워치 SNS 리뷰 데이터와 오피니언 마이닝을 통한 감성 분석 처리에 대한 연구)

  • Shin, Donghyun;Choi, YongLak
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2015.10a
    • /
    • pp.1047-1050
    • /
    • 2015
  • Wearable device, along with IoT(Internet of Things), is considered the core of upcoming generation's convergence technology. Companies are intensely competing one another for prior occupation in the smartwatch market. Consumers that use smartwatch express their preferences by sharing their opinions through SNS(Social Networking Service). Through this study, emotions dictionary is built, which consists of attributes and emotional words related to smartwatch. Based on the emotions dictionary, SNS data has been categorized according to the attributes through opinion data model. Afterwards, overall polarity and attribute polarity of collected data are distinguished through natural language parsing, followed by an analysis of smartwatch reviews. This study will contribute to determination of which attributes of smartwatch to be improved, to arise consumer's interest for individual smartwatch.

  • PDF

Design and Implementation of 'Sea Map' Data Importer Module ('바다지도' 데이터 입력 모듈 설계 및 구현)

  • Yeo, Jimin;Park, Daewon;Park, Suhyun
    • Journal of the Korea Society of Computer and Information
    • /
    • v.19 no.2
    • /
    • pp.91-99
    • /
    • 2014
  • This paper is about design and implementation of 'SeaMap' importer module which is for map-based application using 'SeaMap' data. 'SeaMap' data importer is a module that reads 'SeaMap' data in consistent form and offers using easily change the format and the internal data structure in the application. Design of data importer module is based on analyzing characteristic of 'SeaMap'. The comparative analysis between the data type of 'SeaMap' and standard S-57 Electronic Navigation Chart (ENC) of the International Hydrographic Organization (IHO), based on this, to be applicable of S-52 standards 'SeaMap' data is defined as a structure of data. The importer module is designed and converted to allow to use distribution type of 'SeaMap' data sets in map application, parsing 'SeaMap' data around the object defining transformation data structure. In addition, we implemented a 'SeaMap' data viewer in order to test our 'SeaMap' data importer module.

Eine Wortsyntaktische Analyse der deutschen Ableitung durch QPATR (QPATR를 이용한 독일어 파생어의 어휘구조 분석 - 통합에 기반한 어휘통사론적 접근 -)

  • Choi Byung-Jin
    • Koreanishche Zeitschrift fur Deutsche Sprachwissenschaft
    • /
    • v.3
    • /
    • pp.303-325
    • /
    • 2001
  • In diesem Aufsatz wird versucht, die Wortbildungsstruktur der deutschen Ableitungen zu analysieren und in QPATR zu implementieren. $Daf\"{u}r$ werden die deutschen Ableitungen in morphosyntaktischer Hinsicht betrachtet. Die Wortbildungsstruktur der Ableitungen wird durch Phrasenstrukturregeln mit den detaillierten Unterspezifikationen beschrieben, die durch Unifikation definiert sind. Es wird davon ausgegangen, $da{\ss}\;S\"{a}tze$ durch Phrasenstrukturregeln, die syntaktische Kombinierbarkeit der unmittelbaren Konstituenten definieren, zu analysieren sind. In Wortern sind auch die Regeln $f\"{u}r$ die Erzeugung und Analyse von abgeleiteten $W\"{o}rtern$ zu finden. Die Wortbildungsregln und die Phrasenstrukturregeln von $S\"{a}tzen$ haben $\"{A}hnlichkeit$ von der Form her. $Au{\ss}erdem$ kann man die mophosyntaktischen $Ph\"{a}nomene$ der Kongruenz oder die Selektionsrestriktion nicht nur auf der Satzsyntax, sondern auch auf der Wortsyntax betrachten. D.h. wenn man durch einen Parser mit Hilfe der Phrasenstrukturregeln einen Satz analysieren kann, kann man auch ein Wort aufgrund der Wortbildungsregeln parsen. Im ersten Abschnitt werden die grundlegenden Begriffe der 'unifikationsbasierten Grammatiken', $n\"{a}mlich$ die Begriffe 'Merkmalstruktur', 'Unifikation', und 'Subsumption' $eingef\"{u}hrt$. Im zweiten Abschnitt wird der QPATR - Parser vorgestellt, der aufgrund des 'PATR-II'-Formalismus in $D\"{u}sseldorf$ entwickelt worden ist. Im dritten Abschnitt werden die deutschen abgeleiteten $W\"{o}rter$ durch QPATR analysiert. Bei der Analyse der Worter wird gezeigt, wie die Left-corner-parsing-Strategie funktioniert, und Wie die spezifizierten Informationen durch Unifikation als 'Constraints' die zu kombinierenden unmittelbaren Konstituenten $beschr\"{a}nken$. Zum $Schlu{\ss}$ werden die Ergebnisse der Analyse durch Baumstruktur und Merkmalstruktur $repr\"{a}sentiert$, die durch QPATR automatisch erzeugt werden.

  • PDF

Video-Dissolve Detection using Characteristics of Neighboring Scenes (이웃 장면들의 특성을 이용한 비디오 디졸브 검출)

  • 원종운;최재각;박철현;김범수;곽동민;오상근;박길흠
    • Journal of KIISE:Information Networking
    • /
    • v.30 no.4
    • /
    • pp.504-512
    • /
    • 2003
  • In this paper, we propose a new adaptive dissolve detection method based on the analysis of a dissolve modeling error which is the difference between an ideally modeled dissolve curve with no correlation and an actual dissolve curve including a correlation. The proposed dissolve detection method consists of two steps. First, candidate dissolve regions are extracted using the characteristics of a downward convex parabola, then each candidate region is verified based oil the dissolve modeling error. If the dissolve modeling error for a candidate region is less than a threshold defined by the target modeling error with a target correlation, the candidate region is determined as a resolve region with a lower correlation than the target correlation. The threshold is adaptively determined based on the variances between the candidate regions and the target correlation. By considering the correlation between neighbor scenes, the proposed method is able to be a semantic scene-change detector. The proposed method was tested on various types of data and its performance proved to be more accurate and reliable regardless of variation of variance of test sequences when compared with other commonly use methods.

3D Clothes Modeling of Virtual Human for Metaverse (메타버스를 위한 가상 휴먼의 3차원 의상 모델링)

  • Kim, Hyun Woo;Kim, Dong Eon;Kim, Yujin;Park, In Kyu
    • Journal of Broadcast Engineering
    • /
    • v.27 no.5
    • /
    • pp.638-653
    • /
    • 2022
  • In this paper, we propose the new method of creating 3D virtual-human reflecting the pattern of clothes worn by the person in the high-resolution whole body front image and the body shape data about the person. To get the pattern of clothes, we proceed Instance Segmentation and clothes parsing using Cascade Mask R-CNN. After, we use Pix2Pix to blur the boundaries and estimate the background color and can get UV-Map of 3D clothes mesh proceeding UV-Map base warping. Also, we get the body shape data using SMPL-X and deform the original clothes and body mesh. With UV-Map of clothes and deformed clothes and body mesh, user finally can see the animation of 3D virtual-human reflecting user's appearance by rendering with the state-of-the game engine, i.e. Unreal Engine.

Development of a distributed high-speed data acquisition and monitoring system based on a special data packet format for HUST RF negative ion source

  • Li, Dong;Yin, Ling;Wang, Sai;Zuo, Chen;Chen, Dezhi
    • Nuclear Engineering and Technology
    • /
    • v.54 no.10
    • /
    • pp.3587-3594
    • /
    • 2022
  • A distributed high-speed data acquisition and monitoring system for the RF negative ion source at Huazhong University of Science and Technology (HUST) is developed, which consists of data acquisition, data forwarding and data processing. Firstly, the data acquisition modules sample physical signals at high speed and upload the sampling data with corresponding absolute-time labels over UDP, which builds the time correlation among different signals. And a special data packet format is proposed for the data upload, which is convenient for packing or parsing a fixed-length packet, especially when the span of the time labels in a packet crosses an absolute second. The data forwarding modules then receive the UDP messages and distribute their data packets to the real-time display module and the data storage modules by PUB/SUB-pattern message queue of ZeroMQ. As for the data storage, a scheme combining the file server and MySQL database is adopted to increase the storage rate and facilitate the data query. The test results show that the loss rate of the data packets is within the range of 0-5% and the storage rate is higher than 20 Mbps, both acceptable for the HUST RF negative ion source.

Syntactic and Semantic Disambiguation for Interpretation of Numerals in the Information Retrieval (정보 검색을 위한 숫자의 해석에 관한 구문적.의미적 판별 기법)

  • Moon, Yoo-Jin
    • Journal of the Korea Society of Computer and Information
    • /
    • v.14 no.8
    • /
    • pp.65-71
    • /
    • 2009
  • Natural language processing is necessary in order to efficiently perform filtering tremendous information produced in information retrieval of world wide web. This paper suggested an algorithm for meaning of numerals in the text. The algorithm for meaning of numerals utilized context-free grammars with the chart parsing technique, interpreted affixes connected with the numerals and was designed to disambiguate their meanings systematically supported by the n-gram based words. And the algorithm was designed to use POS (part-of-speech) taggers, to automatically recognize restriction conditions of trigram words, and to gradually disambiguate the meaning of the numerals. This research performed experiment for the suggested system of the numeral interpretation. The result showed that the frequency-proportional method recognized the numerals with 86.3% accuracy and the condition-proportional method with 82.8% accuracy.

Analyzing dependency of Korean subordinate clauses using a composit kernel (복합 커널을 사용한 한국어 종속절의 의존관계 분석)

  • Kim, Sang-Soo;Park, Seong-Bae;Park, Se-Young;Lee, Sang-Jo
    • Korean Journal of Cognitive Science
    • /
    • v.19 no.1
    • /
    • pp.1-15
    • /
    • 2008
  • Analyzing of dependency relation among clauses is one of the most critical parts in parsing Korean sentences because it generates severe ambiguities. To get successful results of analyzing dependency relation, this task has been the target of various machine learning methods including SVM. Especially, kernel methods are usually used to analyze dependency relation and it is reported that they show high performance. This paper proposes an expression and a composit kernel for dependency analysis of Korean clauses. The proposed expression adopts a composite kernel to obtain the similarity among clauses. The composite kernel consists of a parse tree kernel and a liner kernel. A parse tree kernel is used for treating structure information and a liner kernel is applied for using lexical information. the proposed expression is defined as three types. One is a expression of layers in clause, another is relation expression between clause and the other is an expression of inner clause. The experiment is processed by two steps that first is a relation expression between clauses and the second is a expression of inner clauses. The experimental results show that the proposed expression achieves 83.31% of accuracy.

  • PDF

Korean Syntactic Analysis by Using Clausal Segmentation of Embedded Clause (내포문의 단문 분할을 이용한 한국어 구문 분석)

  • Lee, Hyeon-Yeong;Lee, Yong-Seok
    • Journal of KIISE:Software and Applications
    • /
    • v.35 no.1
    • /
    • pp.50-58
    • /
    • 2008
  • Most of Korean sentences are complex sentences which consisted of main clause and embedded clause. These complex sentences have more than one predicate and this causes various syntactic ambiguities in syntactic analysis. These ambiguities are caused by phrase attachment problems which are occurred by the modifying scope of embedded clause. To resolve it, we decide the scope of embedded clause in the sentence and consider this clause as a unit of syntactic category. In this paper, we use sentence patterns information(SPI) and syntactic properties of Korean to decide a scope of embedded clause. First, we split the complex sentence into embedded clause and main clause by the method that embedded clause must have maximal arguments. This work is done by the SPI of the predicate in the embedded clause. And then, the role of this embedded clause is converted into a noun phrases or adverbial phrases in the main clause by the properties of Korean syntax. By this method, the structure of complex sentence is exchanged into a clause. And some phrases attachment problem, which is mainly caused by the modifying scope, is resolved easily. In this paper, we call this method clausal segmentation for embedded clause. By empirical results of parsing 1000 sentences, we found that our method decreases 88.32% of syntactic ambiguities compared to the method that doesn't use SPI and split the sentence with basic clauses.

PPEditor: Semi-Automatic Annotation Tool for Korean Dependency Structure (PPEditor: 한국어 의존구조 부착을 위한 반자동 말뭉치 구축 도구)

  • Kim Jae-Hoon;Park Eun-Jin
    • The KIPS Transactions:PartB
    • /
    • v.13B no.1 s.104
    • /
    • pp.63-70
    • /
    • 2006
  • In general, a corpus contains lots of linguistic information and is widely used in the field of natural language processing and computational linguistics. The creation of such the corpus, however, is an expensive, labor-intensive and time-consuming work. To alleviate this problem, annotation tools to build corpora with much linguistic information is indispensable. In this paper, we design and implement an annotation tool for establishing a Korean dependency tree-tagged corpus. The most ideal way is to fully automatically create the corpus without annotators' interventions, but as a matter of fact, it is impossible. The proposed tool is semi-automatic like most other annotation tools and is designed to edit errors, which are generated by basic analyzers like part-of-speech tagger and (partial) parser. We also design it to avoid repetitive works while editing the errors and to use it easily and friendly. Using the proposed annotation tool, 10,000 Korean sentences containing over 20 words are annotated with dependency structures. For 2 months, eight annotators have worked every 4 hours a day. We are confident that we can have accurate and consistent annotations as well as reduced labor and time.