• Title/Summary/Keyword: R language

Search Result 499, Processing Time 0.028 seconds

Methodology for Deriving Required Quality of Product Using Analysis of Customer Reviews (사용자 리뷰 분석을 통한 제품 요구품질 도출 방법론)

  • Yerin Yu;Jeongeun Byun;Kuk Jin Bae;Sumin Seo;Younha Kim;Namgyu Kim
    • Journal of Information Technology Applications and Management
    • /
    • v.30 no.2
    • /
    • pp.1-18
    • /
    • 2023
  • Recently, as technology development has accelerated and product life cycles have been shortened, it is necessary to derive key product features from customers in the R&D planning and evaluation stage. More companies want differentiated competitiveness by providing consumer-tailored products based on big data and artificial intelligence technology. To achieve this, the need to correctly grasp the required quality, which is a requirement of consumers, is increasing. However, the existing methods are centered on suppliers or domain experts, so there is a gap from the actual perspective of consumers. In other words, product attributes were defined by suppliers or field experts, but this may not consider consumers' actual perspective. Accordingly, the demand for deriving the product's main attributes through reviews containing consumers' perspectives has recently increased. Therefore, we propose a review data analysis-based required quality methodology containing customer requirements. Specifically, a pre-training language model with a good understanding of Korean reviews was established, consumer intent was correctly identified, and key contents were extracted from the review through a combination of KeyBERT and topic modeling to derive the required quality for each product. RevBERT, a Korean review domain-specific pre-training language model, was established through further pre-training. By comparing the existing pre-training language model KcBERT, we confirmed that RevBERT had a deeper understanding of customer reviews. In addition, all processes other than that of selecting the required quality were linked to the automation process, resulting in the automation of deriving the required quality based on data.

Emotion Analysis of Characters in a Comic from State Diagram via Natural Language-based Requirement Specifications

  • Ye Jin Jin;Ji Hoon Kong;Hyun Seung Son;R. Young Chul Kim
    • International journal of advanced smart convergence
    • /
    • v.13 no.1
    • /
    • pp.92-98
    • /
    • 2024
  • The current software industry has an emerging issue with natural language-based requirement specifications. However, the accuracy of such requirement analysis remains a concern. It is noted that most errors still occur at the requirement specification stage. Defining and analyzing requirements based on natural language has become necessary. To address this issue, the linguistic theories of Chomsky and Fillmore are applied to the analysis of natural language-based requirements. This involves identifying the semantics of morphemes and nouns. Consequently, a mechanism was proposed for extracting object state designs and automatically generating code templates. Building on this mechanism, I suggest generating natural language-based comic images. Utilizing state diagrams, I apply changes to the states of comic characters (protagonists) and extract variations in their expressions. This introduces a novel approach to comic image generation. I anticipate highly productive comic creation by applying software processes to Cartoon ART.

An Alignment based technique for Text Translation between Traditional Chinese and Simplified Chinese

  • Sue J. Ker;Lin, Chun-Hsien
    • Proceedings of the Korean Society for Language and Information Conference
    • /
    • 2002.02a
    • /
    • pp.147-156
    • /
    • 2002
  • Aligned parallel corpora have proved very useful in many natural language processing tasks, including statistical machine translation and word sense disambiguation. In this paper, we describe an alignment technique for extracting transfer mapping from the parallel corpus. During building our system and data collection, we observe that there are three types of translation approaches can be used. We especially focuses on Traditional Chinese and Simplified Chinese text lexical translation and a method for extracting transfer mappings for machine translation.

  • PDF

Production and Perception of English /r/ and /l/ by Korean Learners of English: An Experimental Study

  • Kang, Hyeon-Seok
    • Speech Sciences
    • /
    • v.6
    • /
    • pp.7-24
    • /
    • 1999
  • Eleven Korean learners of English took part in an experiment where the production and perception of English /r/ and /l/ in four different word positions was investigated. Overall the subjects made more errors on /l/ in both production and identification tests. The frequency of the subjects' errors was also sensitive to word positions in which the two English liquids occur. Especially the subjects made noticeably fewer errors in intervocalic medial position. It is suggested that the Korean subjects' acquisitional pattern in this particular case of foreign phone learning can be explained more by language particular 'interference' effects rather than 'universal' acoustic arguments such as those given in Dissosway et a1. (1982) and Sheldon and Strange (1982). The results of the experiment also support the minority position among second language educators that in some cases of non-native phone acquisition, learners' production abilities can be developed earlier than their perceptual abilities.

  • PDF

Development of a Layout Editor for Integrated Circuit Design Using ${\Omega}$/PC Graphics Board (${\Omega}$/PC 그래픽 보오드를 이용한 집적회로 설계용 레이아웃 에디터의 개발)

  • Jeong, Gab-Jung;Jang, Ki-Dong;Chung, Ho-Sun;Lee, Wu-Il
    • Journal of the Korean Institute of Telematics and Electronics
    • /
    • v.26 no.5
    • /
    • pp.99-107
    • /
    • 1989
  • This paper describes the KUI-LED (Kyungpook national University Intelligent CAD-Layout EDitor) which is a 2-dimensional graphics editor for IC mask layout. This system runs of IBM PC/AT with the ${\Omega}$/PC graphics board. It offers a sufficient set of facilities to do most kinds of layout. KUIC-LED is written in 'C' and 'Assembly' language.

  • PDF

Distributed Processing of Big Data Analysis based on R using SparkR (SparkR을 이용한 R 기반 빅데이터 분석의 분산 처리)

  • Ryu, Woo-Seok
    • The Journal of the Korea institute of electronic communication sciences
    • /
    • v.17 no.1
    • /
    • pp.161-166
    • /
    • 2022
  • In this paper, we analyze the problems that occur when performing the big data analysis using R as a data analysis tool, and present the usefulness of the data analysis with SparkR which connects R and Spark to support distributed processing of big data effectively. First, we study the memory allocation problem of R which occurs when loading large amounts of data and performing operations, and the characteristics and programming environment of SparkR. And then, we perform the comparison analysis of the execution performance when linear regression analysis is performed in each environment. As a result of the analysis, it was shown that R can be used for data analysis through SparkR without additional language learning, and the code written in R can be effectively processed distributedly according to the increase in the number of nodes in the cluster.

R2RML Based ShEx Schema

  • Choi, Ji-Woong
    • Journal of the Korea Society of Computer and Information
    • /
    • v.23 no.10
    • /
    • pp.45-55
    • /
    • 2018
  • R2RML is a W3C standard language that defines how to expose the relational data as RDF triples. The output from an R2RML mapping is only an RDF dataset. By definition, the dataset has no schema. The lack of schema makes the dataset in linked data portal impractical for integrating and analyzing data. To address this issue, we propose an approach for generating automatically schemas for RDF graphs populated by R2RML mappings. More precisely, we represent the schema using ShEx, which is a language for validating and describing RDF. Our approach allows to generate ShEx schemas as well as RDF datasets from R2RML mappings. Our ShEx schema can provide benefits for both data providers and ordinary users. Data providers can verify and guarantee the structural integrity of the dataset against the schema. Users can write SPARQL queries efficiently by referring to the schema. In this paper, we describe data structures and algorithms of the system to derive ShEx documents from R2RML documents and presents a brief demonstration regarding its proper use.

Distributed/parallel Algorithm Simulator (분산 및 병렬 알고리즘 시뮬레이터)

  • ;R.S.Ramakrishna
    • Proceedings of the Korean Information Science Society Conference
    • /
    • 1999.10c
    • /
    • pp.777-779
    • /
    • 1999
  • A new distributed/parallel algorithm simulator, DASim(Distributed Algorithm Simulator), is proposed in this paper. The idea is to ease the task of design, analysis and implementation of distributed algorithms. A small high level language has been proposed for the purpose. Through this non-language specific high level language, the users are spared from the tedious details about how to program distributed or parallel algorithms. Further, visualization of these algorithms are pretty helpful to understand behaviors of these algorithms.

  • PDF

Deep Lexical Semantics: The Ontological Ascent

  • Hobbs, Jerry R.
    • Proceedings of the Korean Society for Language and Information Conference
    • /
    • 2007.11a
    • /
    • pp.29-41
    • /
    • 2007
  • Concepts of greater and greater complexity can be constructed by building systems of entities, by relating other entities to that system with a figure-ground relation, by embedding concepts of figure-ground in the concept of change, by embedding that in causality, and by coarsening the granularity and beginning the process over again. This process can be called the Ontological Ascent. It pervades natural language discourse, and suggests that to do lexical semantics properly, we must carefully axiomatize abstract theories of systems of entities, the figure-ground relation, change, causality, and granularity. In this paper, I outline what these theories should look like.

  • PDF