A Computational Model of Language Learning Driven by Training Inputs

  • Lee, Eun-Seok (Program of Cognitive Sciences, Seoul National University) ;
  • Lee, Ji-Hoon (Program of Bioinformatics, Seoul National University) ;
  • Zhang, Byoung-Tak (School of Computer Science and Engineering, Seoul National University)
  • Published : 2010.05.28

Abstract

Language learning involves linguistic environments around the learner. So the variation in training input to which the learner is exposed has been linked to their language learning. We explore how linguistic experiences can cause differences in learning linguistic structural features, as investigate in a probabilistic graphical model. We manipulate the amounts of training input, composed of natural linguistic data from animation videos for children, from holistic (one-word expression) to compositional (two- to six-word one) gradually. The recognition and generation of sentences are a "probabilistic" constraint satisfaction process which is based on massively parallel DNA chemistry. Random sentence generation tasks succeed when networks begin with limited sentential lengths and vocabulary sizes and gradually expand with larger ones, like children's cognitive development in learning. This model supports the suggestion that variations in early linguistic environments with developmental steps may be useful for facilitating language acquisition.

Keywords