• 제목/요약/키워드: R language

검색결과 499건 처리시간 0.024초

R 프로그래밍: 통계 계산과 데이터 시각화를 위한 환경 (R programming: Language and Environment for Statistical Computing and Data Visualization)

  • 이두호
    • 전자통신동향분석
    • /
    • 제28권1호
    • /
    • pp.42-51
    • /
    • 2013
  • The R language is an open source programming language and a software environment for statistical computing and data visualization. The R language is widely used among a lot of statisticians and data scientists to develop statistical software and data analysis. The R language provides a variety of statistical and graphical techniques, including basic descriptive statistics, linear or nonlinear modeling, conventional or advanced statistical tests, time series analysis, clustering, simulation, and others. In this paper, we first introduce the R language and investigate its features as a data analytics tool. As results, we may explore the application possibility of the R language in the field of data analytics.

  • PDF

R 언어 기반의 REST API 구현 및 보안문제의 해결 방안 (Implementation of R-language-based REST API and Solution for Security Issues)

  • 강동훈;오세종
    • 예술인문사회 융합 멀티미디어 논문지
    • /
    • 제9권1호
    • /
    • pp.387-394
    • /
    • 2019
  • 최근 빅 데이터의 중요성이 부각되면서 데이터 분석에 대한 수요가 증가하고 있다. R 언어는 데이터 분석을 목적으로 고안된 언어로서, 사용자들은 R언어의 다양한 통계, 머신러닝, 데이터 마이닝 패키지의 알고리즘을 활용하여 데이터를 효과적으로 분석 할 수 있다. 그러나 R 언어는 분석 결과를 어플리케이션으로 만들어 활용하기 어렵다는 단점이 있다. 이를 보완하기 위해 PHP, Java등과 같은 다른 언어를 통해 R 스크립트를 호출하는 법이 제안되었다. 그러나 이러한 개발 방식은 R 이외에도 다른 언어를 혼용해서 작성해야 하는 번거로움이 있다. 본 연구에서는 R 언어의 Plumber 패키지를 활용하여 다른 언어를 사용하지 않고 오직 R 언어만을 이용하여 API를 작성하는 방법을 제안하였다. 또한 API를 구현할 때 고려해야할 보안 이슈와 해결 방안에 대해서도 제시하였다. 본 연구에서 제안한 기술을 이용하여 웹 응용 프로그램을 개발 한다면 높은 생산성과 개발의 편리성, 운영의 효율성을 기대할 수 있다.

R: AN OVERVIEW AND SOME CURRENT DIRECTIONS

  • Tierney, Luke
    • Journal of the Korean Statistical Society
    • /
    • 제36권1호
    • /
    • pp.31-55
    • /
    • 2007
  • R is an open source language for statistical computing and graphics based on the ACM software award-winning S language. R is widely used for data analysis and has become a major vehicle for making available new statistical methodology. This paper presents an overview of the design philosophy and the development model for R, reviews the basic capabilities of the system, and outlines some current projects that will influence future developments of R.

스웨덴어 발음 교육상의 몇 가지 문제점 - 모음을 중심으로 -

  • 변광수
    • 대한음성학회지:말소리
    • /
    • 제4호
    • /
    • pp.20-30
    • /
    • 1982
  • The aim of this paper is to analyse difficulties of the pronunciation in swedish vowels encountered by Koreans learners and to seek solutions in order to correct the possible errors. In the course of the analysis the swedish and Korean vowels in question are compared with the purpose of describing differences aha similarities between these two systems. This contrastive description is largely based on the students' articulatory speech level ana the writer's auditory , judgement . The following points are discussed : 1 ) Vowel length as a distinctive feature in Swedish compared with that of Korean. 2) A special attention is paid on the Swedish vowel [w:] that is characterized by its peculiar type of lip rounding. 3) The six pairs of Swedish vowels that are phonologically contrastive but difficult for Koreans to distinguish one from the other: [y:] ~ [w:], [i:] ~ [y:], [e:] ~ [${\phi}$:], [w;] ~ [u:] [w:] ~ [$\theta$], [$\theta$] ~ [u] 4) The r-colored vowel in the case of the postvocalic /r/ that is very common in American English is not allowed in English sound sequences. The r-colored vowel in the American English pattern has to be broken up and replaced hi-segmental vowel-consonant sequences . Korean accustomed to the American pronunciation are warned in this respect. For a more distinct articulation of the postvocalic /r/ trill [r] is preferred to fricative [z]. 5) The front vowels [e, $\varepsilon, {\;}{\phi}$) become opener variants (${\ae}, {\;}:{\ae}$] before / r / or supradentals. The results of the analysis show that difficulties of the pronunciation of the target language (Swedish) are mostly due to the interference from the Learner's source language (Korean). However, the Learner sometimes tends to get interference also from the other foreign language with which he or she is already familiar when he or she finds in that language more similarity to the target language than in his or her own mother tongue. Hence this foreign language (American English) in this case functions as a second language for Koreans in Learning Swedish.

  • PDF

오픈소스 소프트웨어를 활용한 자연어 처리 패키지 제작에 관한 연구 (Research on Natural Language Processing Package using Open Source Software)

  • 이종화;이현규
    • 한국정보시스템학회지:정보시스템연구
    • /
    • 제25권4호
    • /
    • pp.121-139
    • /
    • 2016
  • Purpose In this study, we propose the special purposed R package named ""new_Noun()" to process nonstandard texts appeared in various social networks. As the Big data is getting interested, R - analysis tool and open source software is also getting more attention in many fields. Design/methodology/approach With more than 9,000 R packages, R provides a user-friendly functions of a variety of data mining, social network analysis and simulation functions such as statistical analysis, classification, prediction, clustering and association analysis. Especially, "KoNLP" - natural language processing package for Korean language - has reduced the time and effort of many researchers. However, as the social data increases, the informal expressions of Hangeul (Korean character) such as emoticons, informal terms and symbols make the difficulties increase in natural language processing. Findings In this study, to solve the these difficulties, special algorithms that upgrade existing open source natural language processing package have been researched. By utilizing the "KoNLP" package and analyzing the main functions in noun extracting command, we developed a new integrated noun processing package "new_Noun()" function to extract nouns which improves more than 29.1% compared with existing package.

추가 사전학습 기반 지식 전이를 통한 국가 R&D 전문 언어모델 구축 (Building Specialized Language Model for National R&D through Knowledge Transfer Based on Further Pre-training)

  • 유은지;서수민;김남규
    • 지식경영연구
    • /
    • 제22권3호
    • /
    • pp.91-106
    • /
    • 2021
  • 최근 딥러닝 기술이 빠르게 발전함에 따라 국가 R&D 분야의 방대한 텍스트 문서를 다양한 관점에서 분석하기 위한 수요가 급증하고 있다. 특히 대용량의 말뭉치에 대해 사전학습을 수행한 BERT(Bidirectional Encoder Representations from Transformers) 언어모델의 활용에 대한 관심이 높아지고 있다. 하지만 국가 R&D와 같이 고도로 전문화된 분야에서 높은 빈도로 사용되는 전문어는 기본 BERT에서 충분히 학습이 이루어지지 않은 경우가 많으며, 이는 BERT를 통한 전문 분야 문서 이해의 한계로 지적되고 있다. 따라서 본 연구에서는 최근 활발하게 연구되고 있는 추가 사전학습을 활용하여, 기본 BERT에 국가 R&D 분야 지식을 전이한 R&D KoBERT 언어모델을 구축하는 방안을 제시한다. 또한 제안 모델의 성능 평가를 위해 보건의료, 정보통신 분야의 과제 약 116,000건을 대상으로 분류 분석을 수행한 결과, 제안 모델이 순수한 KoBERT 모델에 비해 정확도 측면에서 더 높은 성능을 나타내는 것을 확인하였다.

온톨로지 기반 연구개발 평가지표 선정기법 (An Ontological Approach to Select R&D Evaluation Metrics)

  • 이희정
    • 산업경영시스템학회지
    • /
    • 제33권1호
    • /
    • pp.80-90
    • /
    • 2010
  • Performance management is very popular in business area and seems to be an exciting topic. Despite significant research efforts and myriads of performance metrics, performance management today as a rigorous approach is still in an immature state and metrics are often selected based on intuitive and heuristic approach. In a R&D sector, the difficulty to select the proper performance metrics is even more increasing due to the natural characteristics of R&D such as unique or domain-specific problems. In this paper, we present a way of presenting R&D performance framework using ontology language. Based on this, the specific metrics can be derived by reusing or inheriting the context in the framework. The proposed ontological framework is formalized using OWL(Ontology Web Language) and metrics selection rules satisfying the characteristics of R&D are represented in SWRL(Semantic Web Rule Language). Actual metrics selection procedure is carried out using JESS rule engine, a plug-in to Prot$\acute{e}$g$\acute{e}$, and illustrated with an example, incorporating a prevalent R&D performance model : TVP(Technology Value Pyramid).

Language Lateralization in Patients with Temporal Lobe Epilepsy : A Comparison between Volumetric Analysis and the Wada Test

  • Oh, Young-Min;Koh, Eun-Jeong
    • Journal of Korean Neurosurgical Society
    • /
    • 제45권6호
    • /
    • pp.329-335
    • /
    • 2009
  • Objective : Determining language lateralization is important for the presurgical evaluation of patients with medically intractable epilepsy. The Wada test has been the gold standard for lateralization of language dominance before epilepsy surgery. However, it is an invasive test with risk, and have some limitations. Methods : We compared the volumetric analysis with Wada test, and studied the clinical potential of volumetric analysis to assess language laterality in large surgical candidates with temporal lobe epilepsy (TLE). To examine the efficacy of volumetric analysis to determine language lateralization during presurgical evaluation, we compared the volumetric analysis of the bilateral planum temporale with the results of Wada test in 59 patients with chronic intractable TLE (rTLE, n=32; lTLE, n=27) who underwent epilepsy surgery. We measured the gray matter volumes of planum temporale (PT) of each patients using the VoxelPlus2 program (Mevisys, Daejeon, Korea). Results : Overall congruence of the volumetric analysis with the Wada test was 97.75% in rTLE patients and 81.5% in lTLE patients. There were more significant leftward asymmetry of the PT in rTLE patients than lTLE patients. In lTLE patients, relatively high proportion (37%) of the patients showed bilateral or right hemispheric language dominance. Conclusion : These results provide evidence that the volumetric analysis of the PT could be used as an alternatives in language lateralization. Also, the results of the Wada test suggested that there was considerable plasticity of language representation in the brains of patients with intractable TLE and it was associated with an earlier age of brain injury.

Fast Ftatistical Grammar Induction

  • Hogenhout, Wide-R.;Matsumoto, Yuji
    • 한국언어정보학회:학술대회논문집
    • /
    • 한국언어정보학회 1996년도 Language, Information and Computation = Selected Papers from the 11th Pacific Asia Conference on Language, Information and Computation, Seoul
    • /
    • pp.383-392
    • /
    • 1996
  • PDF

딥러닝 사전학습 언어모델 기술 동향 (Recent R&D Trends for Pretrained Language Model)

  • 임준호;김현기;김영길
    • 전자통신동향분석
    • /
    • 제35권3호
    • /
    • pp.9-19
    • /
    • 2020
  • Recently, a technique for applying a deep learning language model pretrained from a large corpus to fine-tuning for each application task has been widely used as a language processing technology. The pretrained language model shows higher performance and satisfactory generalization performance than existing methods. This paper introduces the major research trends related to deep learning pretrained language models in the field of language processing. We describe in detail the motivations, models, learning methods, and results of the BERT language model that had significant influence on subsequent studies. Subsequently, we introduce the results of language model studies after BERT, focusing on SpanBERT, RoBERTa, ALBERT, BART, and ELECTRA. Finally, we introduce the KorBERT pretrained language model, which shows satisfactory performance in Korean language. In addition, we introduce techniques on how to apply the pretrained language model to Korean (agglutinative) language, which consists of a combination of content and functional morphemes, unlike English (refractive) language whose endings change depending on the application.