Browse > Article
http://dx.doi.org/10.3745/KTSDE.2022.11.3.133

Comparison of Korean Classification Models' Korean Essay Score Range Prediction Performance  

Cho, Heeryon (중앙대학교 인문콘텐츠연구소)
Im, Hyeonyeol (중앙대학교 다빈치교양대학)
Yi, Yumi (중앙대학교 인문콘텐츠연구소)
Cha, Junwoo (중앙대학교 한국어교육원)
Publication Information
KIPS Transactions on Software and Data Engineering / v.11, no.3, 2022 , pp. 133-140 More about this Journal
Abstract
We investigate the performance of deep learning-based Korean language models on a task of predicting the score range of Korean essays written by foreign students. We construct a data set containing a total of 304 essays, which include essays discussing the criteria for choosing a job ('job'), conditions of a happy life ('happ'), relationship between money and happiness ('econ'), and definition of success ('succ'). These essays were labeled according to four letter grades (A, B, C, and D), and a total of eleven essay score range prediction experiments were conducted (i.e., five for predicting the score range of 'job' essays, five for predicting the score range of 'happiness' essays, and one for predicting the score range of mixed topic essays). Three deep learning-based Korean language models, KoBERT, KcBERT, and KR-BERT, were fine-tuned using various training data. Moreover, two traditional probabilistic machine learning classifiers, naive Bayes and logistic regression, were also evaluated. Experiment results show that deep learning-based Korean language models performed better than the two traditional classifiers, with KR-BERT performing the best with 55.83% overall average prediction accuracy. A close second was KcBERT (55.77%) followed by KoBERT (54.91%). The performances of naive Bayes and logistic regression classifiers were 52.52% and 50.28% respectively. Due to the scarcity of training data and the imbalance in class distribution, the overall prediction performance was not high for all classifiers. Moreover, the classifiers' vocabulary did not explicitly capture the error features that were helpful in correctly grading the Korean essay. By overcoming these two limitations, we expect the score range prediction performance to improve.
Keywords
Deep Learning-Based Korean Language Model; KoBERT; KcBERT; KR-BERT; Document Classification;
Citations & Related Records
Times Cited By KSCI : 4  (Citation Analysis)
연도 인용수 순위
1 S. S. Kang and E. S. Jang, "Automatic scoring system for Korean short answers by student answer analysis and answer template construction," KIISE Transactions on Computing Practices, Vol.22, No.5, pp.218-224, 2016.   DOI
2 J. Heo and S.-Y. Park, "Design and implementation of an automatic scoring model using a voting method for descriptive answers," Journal of the Korea Society of Computer and Information, Vol.18, No.8, pp.17-25, 2013.   DOI
3 M.-A. Cheon, C.-H. Kim, J.-H. Kim, E.-H. Noh, K.-H. Sung, and M.-Y. Song, "Automated scoring system for Korean short-answer questions using predictability and unanimity," KIPS Transactions on Software and Data Engineering, Vol.5, No.11, pp.527-534, 2016.   DOI
4 J.-Y. Choi and H.-S. Lim, "E-commerce data based Sentiment analysis model implementation using natural language processing model," Journal of the Korea Convergence Society, Vo.11, No.11, pp.33-39, 2020.   DOI
5 T.-H. Kim, D.-B. Cho, H.-Y. Lee, H.-J. Won, and S.-S. Kang, "Sentiment analysis system by using BERT language model," in Proceedings of the KIPS Spring Conference 2020, Vol.27, No.2, pp.975-977, 2020.
6 S. Park, H. Yang, M. Choe, M. Ha, K. Chung, and M. Koo, "Sentimental analysis of YouTube Korean comments using KoBERT," in Proceedings of Korea Software Congress 2020, pp.1385-1387, 2020.
7 A.-G. Kim and Y.-S. Jeong, "Topic classification of domestic music using KoBERT," in Proceedings of Korea Software Congress 2021, pp.1738-1740, 2021.
8 J. Devlin, M.-W. Chang, K. Lee, and K. Toutanova, "BERT: Pre-training of deep bidirectional transformers for language understanding," in Proceedings of the 2019 Annual Conference of the North American Chapter of the Association for Computational Linguistics, pp.4171-4186, 2019.
9 H. Cho, H. Im, J. Cha, and Y. Yi, "Comparison of automatic score range prediction of Korean essays using KoBERT, Naive Bayes & Logistic Regression," in Proceedings of the KIPS Spring Conference 2021, Vol.28, No.1, pp.501-504, 2021.
10 M. Banko and E. Brill, "Scaling to very very large corpora for natural language disambiguation," in Proceedings of the 39th Annual Meeting of the Association for Computational Linguistics, pp.26-33, 2001.
11 S. Yoo and K. Yang, "The status of Korean as an international language," Hallyu Now: Global Hallyu Issue Magazine, Vol.34, pp.9-16, 2020.
12 H. J. Park and W. S. Kang, "Design and implementation of a subjective-type evaluation system using natural language processing technique," The Journal of Korean Association of Computer Education, Vol.6, No.3, pp.207-216, 2003.
13 I.-N. Park, S.-S. Kang, E.-H. Noh, M.-H. Kim, and T.-J. Seong, "Automatic scoring of Korean short answers by answer template description," Journal of KIISE: Computing Practices and Letters, Vol.19, No.12, pp.630-636, 2013.
14 J. Lee, "KcBERT: Korean comments BERT," in Proceedings of the 32nd Annual Conference on Human and Cognitive Language Technology, pp.437-440, 2020.
15 Y.-J. Lee and H.-J. Choi, "Joint Learning-based KoBERT for emotion recognition in Korean," in Proceedings of Korea Software Congress 2020, pp.568-570, 2020.
16 K. H. Park and Y.-S. Jeong, "Korean daily conversation topics classification using KoBERT," in Proceedings of Korea Software Congress 2021, pp.1735-1737, 2021.
17 M.-A. Cheon, H.-W. Seo, J.-H. Kim, E.-H. Noh, K.-H. Sung, and E. Young Lim, "Semi-automatic scoring for short Korean free-text responses using semi-supervised learning," Korean Journal of Cognitive Science, Vol.26, No.2, pp.147-165, 2015.   DOI
18 S. Hwang and D. Kim, "BERT-based classification model for Korean documents," The Journal of Society for e-Business Studies, Vol.25, No.1, pp.203-214, 2020.   DOI
19 H. Cho, Y. Yi, H. Im, J. Cha, and C. Lee, "Automatic score range classification of Korean essays using deep learning-based Korean language models -The case of KoBERT & KoGPT2-," Journal of the International Network for Korean Language and Culture, Vol.18, No.1, pp.217-241, 2021.   DOI
20 S. Lee, H. Jang, Y. Baik, S. Park, and H. Shin, "KR-BERT: A small-scale Korean-specific language model," ArXiv, 2020. [Internet], https://arxiv.org/abs/2008.03979.