• Title/Summary/Keyword: Language model

Search Result 2,772, Processing Time 0.026 seconds

A Study on Teaching Methods of Mathematics Using SIOP Model for KLLs (SIOP 모델을 적용한 한국어학습자의 수학 학습 지도 방안 연구)

  • Choi, Hee Hoon;Chang, Hyewon
    • Journal of Elementary Mathematics Education in Korea
    • /
    • v.23 no.3
    • /
    • pp.305-321
    • /
    • 2019
  • Rapid demographic changes such as international marriages and immigration have led to the transition of Korea to a multicultural society, thereby causing the need for education for multicultural students. In particular, there is a growing need to support Korean Language Learners (KLLs) who learn in Korean in their classrooms and whose native language is a foreign language. This study aims to adapt some teaching strategies of the SIOP model developed in the U.S. for English Language Learners(ELLs) to fit classroom situations in Korea and apply them to the Korean language learners to analyze the features of mathematical communication and to examine the possibility of a change in mathematical errors. Specifically, three KLLs in 5th grade participated in seven geometry lessons adapting some characteristics of SIOP model and then, their mathematical communication and mathematical errors were analyzed. The results of this study are expected to provide didactical implications for identifying characteristics of KLLs and for setting direction for teaching them mathematics.

  • PDF

A Unicode based Deep Handwritten Character Recognition model for Telugu to English Language Translation

  • BV Subba Rao;J. Nageswara Rao;Bandi Vamsi;Venkata Nagaraju Thatha;Katta Subba Rao
    • International Journal of Computer Science & Network Security
    • /
    • v.24 no.2
    • /
    • pp.101-112
    • /
    • 2024
  • Telugu language is considered as fourth most used language in India especially in the regions of Andhra Pradesh, Telangana, Karnataka etc. In international recognized countries also, Telugu is widely growing spoken language. This language comprises of different dependent and independent vowels, consonants and digits. In this aspect, the enhancement of Telugu Handwritten Character Recognition (HCR) has not been propagated. HCR is a neural network technique of converting a documented image to edited text one which can be used for many other applications. This reduces time and effort without starting over from the beginning every time. In this work, a Unicode based Handwritten Character Recognition(U-HCR) is developed for translating the handwritten Telugu characters into English language. With the use of Centre of Gravity (CG) in our model we can easily divide a compound character into individual character with the help of Unicode values. For training this model, we have used both online and offline Telugu character datasets. To extract the features in the scanned image we used convolutional neural network along with Machine Learning classifiers like Random Forest and Support Vector Machine. Stochastic Gradient Descent (SGD), Root Mean Square Propagation (RMS-P) and Adaptative Moment Estimation (ADAM)optimizers are used in this work to enhance the performance of U-HCR and to reduce the loss function value. This loss value reduction can be possible with optimizers by using CNN. In both online and offline datasets, proposed model showed promising results by maintaining the accuracies with 90.28% for SGD, 96.97% for RMS-P and 93.57% for ADAM respectively.

Design of a Korean Speech Recognition Platform (한국어 음성인식 플랫폼의 설계)

  • Kwon Oh-Wook;Kim Hoi-Rin;Yoo Changdong;Kim Bong-Wan;Lee Yong-Ju
    • MALSORI
    • /
    • no.51
    • /
    • pp.151-165
    • /
    • 2004
  • For educational and research purposes, a Korean speech recognition platform is designed. It is based on an object-oriented architecture and can be easily modified so that researchers can readily evaluate the performance of a recognition algorithm of interest. This platform will save development time for many who are interested in speech recognition. The platform includes the following modules: Noise reduction, end-point detection, met-frequency cepstral coefficient (MFCC) and perceptually linear prediction (PLP)-based feature extraction, hidden Markov model (HMM)-based acoustic modeling, n-gram language modeling, n-best search, and Korean language processing. The decoder of the platform can handle both lexical search trees for large vocabulary speech recognition and finite-state networks for small-to-medium vocabulary speech recognition. It performs word-dependent n-best search algorithm with a bigram language model in the first forward search stage and then extracts a word lattice and restores each lattice path with a trigram language model in the second stage.

  • PDF

A Language Model Approach to "The Vegetarian" (채식주의자: 랭귀지 모델 접근)

  • Kim, Jaejun;Kwon, Junhyeok;Kim, Yoolae;Park, Myung-Kwan;Song, Sanghoun
    • 한국어정보학회:학술대회논문집
    • /
    • 2017.10a
    • /
    • pp.260-263
    • /
    • 2017
  • This paper is to broaden the possible spectrums of analyzing the Korean-written novel "The Vegetarian" by using the computational linguistics program. Through the use of language model, which was usually used in bi-gram analysis in corpus linguistics, to the International Man Booker award winning novel, the characteristics of "The Vegetarian" is investigated by comparing it to the English-written novel "A Little Life".

  • PDF

A Computational Model of Language Learning Driven by Training Inputs

  • Lee, Eun-Seok;Lee, Ji-Hoon;Zhang, Byoung-Tak
    • Proceedings of the Korean Society for Cognitive Science Conference
    • /
    • 2010.05a
    • /
    • pp.60-65
    • /
    • 2010
  • Language learning involves linguistic environments around the learner. So the variation in training input to which the learner is exposed has been linked to their language learning. We explore how linguistic experiences can cause differences in learning linguistic structural features, as investigate in a probabilistic graphical model. We manipulate the amounts of training input, composed of natural linguistic data from animation videos for children, from holistic (one-word expression) to compositional (two- to six-word one) gradually. The recognition and generation of sentences are a "probabilistic" constraint satisfaction process which is based on massively parallel DNA chemistry. Random sentence generation tasks succeed when networks begin with limited sentential lengths and vocabulary sizes and gradually expand with larger ones, like children's cognitive development in learning. This model supports the suggestion that variations in early linguistic environments with developmental steps may be useful for facilitating language acquisition.

  • PDF

Study on Various Factors Associated with the Effects of Cyber Home Study in Korean Language Education based on Structural Equation Model (구조방정식을 이용한 국어 사이버 가정학습의 효과 관련 요인에 관한 연구)

  • Lim, Mi-Ja;Baek, Hyeon-Gi
    • Journal of Digital Convergence
    • /
    • v.6 no.1
    • /
    • pp.83-91
    • /
    • 2008
  • The objective of this research is to assess various factors affecting E-learning in Korean language education. In this research, we hypothesize that several factors such as absorption, motivation and tutors increase the educational effects of E-learning and ultimately affect learning attitude and satisfaction of students in E-learning. To discuss the hypothesis, we analyzed survey data of 300 students who were taking E-learning class of Korean language for three weeks in October 2007 based on Structural Equation Model. The result of our analysis shows that the factors such as absorption, motivation, tutors have positive effects on E-learning in Korean language education and positive influence on learning attitude and satisfaction on students as well.

  • PDF

A Semi-supervised Learning of HMM to Build a POS Tagger for a Low Resourced Language

  • Pattnaik, Sagarika;Nayak, Ajit Kumar;Patnaik, Srikanta
    • Journal of information and communication convergence engineering
    • /
    • v.18 no.4
    • /
    • pp.207-215
    • /
    • 2020
  • Part of speech (POS) tagging is an indispensable part of major NLP models. Its progress can be perceived on number of languages around the globe especially with respect to European languages. But considering Indian Languages, it has not got a major breakthrough due lack of supporting tools and resources. Particularly for Odia language it has not marked its dominancy yet. With a motive to make the language Odia fit into different NLP operations, this paper makes an attempt to develop a POS tagger for the said language on a HMM (Hidden Markov Model) platform. The tagger judiciously considers bigram HMM with dynamic Viterbi algorithm to give an output annotated text with maximum accuracy. The model is experimented on a corpus belonging to tourism domain accounting to a size of approximately 0.2 million tokens. With the proportion of training and testing as 3:1, the proposed model exhibits satisfactory result irrespective of limited training size.

Automatic Coarticulation Detection for Continuous Sign Language Recognition (연속된 수화 인식을 위한 자동화된 Coarticulation 검출)

  • Yang, Hee-Deok;Lee, Seong-Whan
    • Journal of KIISE:Software and Applications
    • /
    • v.36 no.1
    • /
    • pp.82-91
    • /
    • 2009
  • Sign language spotting is the task of detecting and recognizing the signs in a signed utterance. The difficulty of sign language spotting is that the occurrences of signs vary in both motion and shape. Moreover, the signs appear within a continuous gesture stream, interspersed with transitional movements between signs in a vocabulary and non-sign patterns(which include out-of-vocabulary signs, epentheses, and other movements that do not correspond to signs). In this paper, a novel method for designing a threshold model in a conditional random field(CRF) model is proposed. The proposed model performs an adaptive threshold for distinguishing between signs in the vocabulary and non-sign patterns. A hand appearance-based sign verification method, a short-sign detector, and a subsign reasoning method are included to further improve sign language spotting accuracy. Experimental results show that the proposed method can detect signs from continuous data with an 88% spotting rate and can recognize signs from isolated data with a 94% recognition rate, versus 74% and 90% respectively for CRFs without a threshold model, short-sign detector, subsign reasoning, and hand appearance-based sign verification.

Design and Implementation of XML Based Relational Database Metadata Repository (XML을 기반으로 한 관계형 데이터베이스 메타데이터 리파지토리 설계 및 구현)

  • Gwon, Eun-Jeong;Yong, Hwan-Seung
    • The KIPS Transactions:PartD
    • /
    • v.9D no.1
    • /
    • pp.1-10
    • /
    • 2002
  • Metadata is data about data that is used to mange data itself. As applications based on DBMS are increased, it is suggested that metadata model and metadata interchange model to manage metadata in DBMS. but metadata which is in the form of XML (eXtensible Markup Language) document is generally stored into RDBMS. Therefore In this paper, as for the method to store metadata of RDBMS into OODBMS, we design metadata model, metadata interchange model and implement new repository system. The metadata of RDBMS is translated into in the form of XML Document and integrated into XML Data Server on OODBMS, eXcelon and executes retrieval metadata information about RDBMS by XQL(XML Query Language). So It is possible to searchm edit a metadata. The metadata of XML documents stored in eXcelon is easily made to be printed in web browser by applying a XSL (extensible StyleSheets Language). So we can have a detail information about property of metadata in DBMS.

Language Model based on VCCV and Test of Smoothing Techniques for Sentence Speech Recognition (문장음성인식을 위한 VCCV 기반의 언어모델과 Smoothing 기법 평가)

  • Park, Seon-Hee;Roh, Yong-Wan;Hong, Kwang-Seok
    • The KIPS Transactions:PartB
    • /
    • v.11B no.2
    • /
    • pp.241-246
    • /
    • 2004
  • In this paper, we propose VCCV units as a processing unit of language model and compare them with clauses and morphemes of existing processing units. Clauses and morphemes have many vocabulary and high perplexity. But VCCV units have low perplexity because of the small lexicon and the limited vocabulary. The construction of language models needs an issue of the smoothing. The smoothing technique used to better estimate probabilities when there is an insufficient data to estimate probabilities accurately. This paper made a language model of morphemes, clauses and VCCV units and calculated their perplexity. The perplexity of VCCV units is lower than morphemes and clauses units. We constructed the N-grams of VCCV units with low perplexity and tested the language model using Katz, absolute, modified Kneser-Ney smoothing and so on. In the experiment results, the modified Kneser-Ney smoothing is tested proper smoothing technique for VCCV units.