• Title/Summary/Keyword: language representation model

Search Result 139, Processing Time 0.026 seconds

Zero-anaphora resolution in Korean based on deep language representation model: BERT

  • Kim, Youngtae;Ra, Dongyul;Lim, Soojong
    • ETRI Journal
    • /
    • v.43 no.2
    • /
    • pp.299-312
    • /
    • 2021
  • It is necessary to achieve high performance in the task of zero anaphora resolution (ZAR) for completely understanding the texts in Korean, Japanese, Chinese, and various other languages. Deep-learning-based models are being employed for building ZAR systems, owing to the success of deep learning in the recent years. However, the objective of building a high-quality ZAR system is far from being achieved even using these models. To enhance the current ZAR techniques, we fine-tuned a pretrained bidirectional encoder representations from transformers (BERT). Notably, BERT is a general language representation model that enables systems to utilize deep bidirectional contextual information in a natural language text. It extensively exploits the attention mechanism based upon the sequence-transduction model Transformer. In our model, classification is simultaneously performed for all the words in the input word sequence to decide whether each word can be an antecedent. We seek end-to-end learning by disallowing any use of hand-crafted or dependency-parsing features. Experimental results show that compared with other models, our approach can significantly improve the performance of ZAR.

RBM-based distributed representation of language (RBM을 이용한 언어의 분산 표상화)

  • You, Heejo;Nam, Kichun;Nam, Hosung
    • Korean Journal of Cognitive Science
    • /
    • v.28 no.2
    • /
    • pp.111-131
    • /
    • 2017
  • The connectionist model is one approach to studying language processing from a computational perspective. And building a representation in the connectionist model study is just as important as making the structure of the model in that it determines the level of learning and performance of the model. The connectionist model has been constructed in two different ways: localist representation and distributed representation. However, the localist representation used in the previous studies had limitations in that the unit of the output layer having a rare target activation value is inactivated, and the past distributed representation has the limitation of difficulty in confirming the result by the opacity of the displayed information. This has been a limitation of the overall connection model study. In this paper, we present a new method to induce distributed representation with local representation using abstraction of information, which is a feature of restricted Boltzmann machine, with respect to the limitation of such representation of the past. As a result, our proposed method effectively solves the problem of conventional representation by using the method of information compression and inverse transformation of distributed representation into local representation.

An implementation of sign language communication sytem between korean and japanese using 3D model (3차원 모델을 이용한 한-일 수화 영상통신 시스템의 구현)

  • ;;Yoshinao Aoki
    • Proceedings of the IEEK Conference
    • /
    • 1998.06a
    • /
    • pp.925-928
    • /
    • 1998
  • In this paper, we implemented a sign language communication system between korean and japanese using 3D model. Unitl now, we have developed the system through 2D image. The system, however, has some drawbacks base don the limitations of 2D model. Especially it is not comprehensive in the sign language gesture including forward or backward movements becasue it can only display a front view. Therefore, we employed a 3D model for the representation of sign language. The proposed communication system si implemented with windows 95 visual C++ 5.0 imported open inventor library on pentium 233 system. Experimented results show that the 3D system can solve the problems in 2D representation.

  • PDF

Comparative study of text representation and learning for Persian named entity recognition

  • Pour, Mohammad Mahdi Abdollah;Momtazi, Saeedeh
    • ETRI Journal
    • /
    • v.44 no.5
    • /
    • pp.794-804
    • /
    • 2022
  • Transformer models have had a great impact on natural language processing (NLP) in recent years by realizing outstanding and efficient contextualized language models. Recent studies have used transformer-based language models for various NLP tasks, including Persian named entity recognition (NER). However, in complex tasks, for example, NER, it is difficult to determine which contextualized embedding will produce the best representation for the tasks. Considering the lack of comparative studies to investigate the use of different contextualized pretrained models with sequence modeling classifiers, we conducted a comparative study about using different classifiers and embedding models. In this paper, we use different transformer-based language models tuned with different classifiers, and we evaluate these models on the Persian NER task. We perform a comparative analysis to assess the impact of text representation and text classification methods on Persian NER performance. We train and evaluate the models on three different Persian NER datasets, that is, MoNa, Peyma, and Arman. Experimental results demonstrate that XLM-R with a linear layer and conditional random field (CRF) layer exhibited the best performance. This model achieved phrase-based F-measures of 70.04, 86.37, and 79.25 and word-based F scores of 78, 84.02, and 89.73 on the MoNa, Peyma, and Arman datasets, respectively. These results represent state-of-the-art performance on the Persian NER task.

A multilingual grammar model of honorification: using the HPSG and MRS formalism

  • Song, Sanghoun
    • Language and Information
    • /
    • v.20 no.1
    • /
    • pp.25-49
    • /
    • 2016
  • Honorific forms express the speaker's social attitude to others and also indicate the social ranks and level of intimacy of the participants in the discourse. In a cross-linguistic perspective of grammar engineering, modelling honorification has been regarded as a key strategy for improving language processing applications. Using the HPSG and MRS formalism, this article provides a multilingual grammar model of honorification. The present study incorporates the honorific information into the Meaning Representation System (MRS) via Individual Constraints (ICONS), and then conducts an evaluation to see if the model contributes to semantics-based language processing.

  • PDF

Language-based Classification of Words using Deep Learning (딥러닝을 이용한 언어별 단어 분류 기법)

  • Zacharia, Nyambegera Duke;Dahouda, Mwamba Kasongo;Joe, Inwhee
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2021.05a
    • /
    • pp.411-414
    • /
    • 2021
  • One of the elements of technology that has become extremely critical within the field of education today is Deep learning. It has been especially used in the area of natural language processing, with some word-representation vectors playing a critical role. However, some of the low-resource languages, such as Swahili, which is spoken in East and Central Africa, do not fall into this category. Natural Language Processing is a field of artificial intelligence where systems and computational algorithms are built that can automatically understand, analyze, manipulate, and potentially generate human language. After coming to discover that some African languages fail to have a proper representation within language processing, even going so far as to describe them as lower resource languages because of inadequate data for NLP, we decided to study the Swahili language. As it stands currently, language modeling using neural networks requires adequate data to guarantee quality word representation, which is important for natural language processing (NLP) tasks. Most African languages have no data for such processing. The main aim of this project is to recognize and focus on the classification of words in English, Swahili, and Korean with a particular emphasis on the low-resource Swahili language. Finally, we are going to create our own dataset and reprocess the data using Python Script, formulate the syllabic alphabet, and finally develop an English, Swahili, and Korean word analogy dataset.

Simple and effective neural coreference resolution for Korean language

  • Park, Cheoneum;Lim, Joonho;Ryu, Jihee;Kim, Hyunki;Lee, Changki
    • ETRI Journal
    • /
    • v.43 no.6
    • /
    • pp.1038-1048
    • /
    • 2021
  • We propose an end-to-end neural coreference resolution for the Korean language that uses an attention mechanism to point to the same entity. Because Korean is a head-final language, we focused on a method that uses a pointer network based on the head. The key idea is to consider all nouns in the document as candidates based on the head-final characteristics of the Korean language and learn distributions over the referenced entity positions for each noun. Given the recent success of applications using bidirectional encoder representation from transformer (BERT) in natural language-processing tasks, we employed BERT in the proposed model to create word representations based on contextual information. The experimental results indicated that the proposed model achieved state-of-the-art performance in Korean language coreference resolution.

A Spatial Structural Query Language-G/SQL

  • Fang, Yu;Chu, Fang;Xinming, Tang
    • Proceedings of the KSRS Conference
    • /
    • 2002.10a
    • /
    • pp.860-879
    • /
    • 2002
  • Traditionally, Geographical Information Systems can only process spatial data in a procedure-oriented way, and the data can't be treated integrally. This method limits the development of spatial data applications. A new and promising method to solve this problem is the spatial structural query language, which extends SQL and provides integrated accessing to spatial data. In this paper, the theory of spatial structural query language is discussed, and a new geographical data model based on the concepts and data model in OGIS is introduced. According to this model, we implemented a spatial structural query language G/SQL. Through the studies of the 9-Intersection Model, G/SQL provides a set of topological relational predicates and spatial functions for GIS application development. We have successfully developed a Web-based GIS system-WebGIS-using G/SQL. Experiences show that the spatial operators G/SQL offered are complete and easy-to-use. The BNF representation of G/SQL syntax is included in this paper.

  • PDF

A Structured Markup Language for the Object-Oriented Representation and Management of Decision Models on the Web (웹상에서의 의사결정모형의 객체지향적 표현과 관리를 위한 구조적 마크업 언어)

  • Kim, Hyoung-Do
    • Asia pacific journal of information systems
    • /
    • v.8 no.2
    • /
    • pp.53-67
    • /
    • 1998
  • The explosive growth of the Web is providing end-users access to ever-increasing volumes of information. The resources of legacy systems and relational databases have also been made available to the Web browser, which has become an essential business tool. Recently, model management on the Internet/Web is also proposed with its conceptual design or prototypical system like DecisionNet and DSS Web. However, they are also suffering from the same symptoms as the Web, Although we can identify the elements of a page with HTML tags and (declare) the relationships among the various document elements, they are semantically opaque to computer systems and have no domain-specific meaning. However, HTML is not extensible, so developers are forced to invent convoluted, non-standard solutions for embedding and parsing data. Extensible Markup Language (XML) is a simplified subset of SGML that has many benefits for folks who want to improve structure, maintainability, searchability, presentation, and other aspects of their document management. This paper proposes a structured markup language for model representation and management on the Web as an XML application. The language is based on a conceptual modeling framework, Object-Oriented Structured Modeling (OOSM), which is an extension of the structured modeling.

  • PDF

Model Multiplicity (UML) Versus Model Singularity in System Requirements and Design

  • Al-Fedaghi, Sabah
    • International Journal of Computer Science & Network Security
    • /
    • v.21 no.4
    • /
    • pp.103-114
    • /
    • 2021
  • A conceptual model can be used to manage complexity in both the design and implementation phases of the system development life cycle. Such a model requires a firm grasp of the abstract principles on which a system is based, as well as an understanding of the high-level nature of the representation of entities and processes. In this context, models can have distinct architectural characteristics. This paper discusses model multiplicity (e.g., unified modeling language [UML]), model singularity (e.g., object-process methodology [OPM], thinging machine [TM]), and a heterogeneous model that involves multiplicity and singularity. The basic idea of model multiplicity is that it is not possible to present all views in a single representation, so a number of models are used, with each model representing a different view. The model singularity approach uses only a single unified model that assimilates its subsystems into one system. This paper is concerned with current approaches, especially in software engineering texts, where multimodal UML is introduced as the general-purpose modeling language (i.e., UML is modeling). In such a situation, we suggest raising the issue of multiplicity versus singularity in modeling. This would foster a basic appreciation of the UML advantages and difficulties that may be faced during modeling, especially in the educational setting. Furthermore, we advocate the claim that a multiplicity of views does not necessitate a multiplicity of models. The model singularity approach can represent multiple views (static, behavior) without resorting to a collection of multiple models with various notations. We present an example of such a model where the static representation is developed first. Then, the dynamic view and behavioral representations are built by incorporating a decomposition strategy interleaved with the notion of time.