Acknowledgement
This work was funded by the Korea Electric Power Corporation (KEPCO) (R22XO02-30) and "Regional Innovation Strategy (RIS)" through the National Research Foundation of Korea(NRF) funded by the Ministry of Education (MOE) (2021RIS-004).
References
- W. Y. Wang and D. Yang, "That's so annoying!!!: A lexical and frame-semantic embedding based data augmentation approach to automatic categorization of annoying behaviors using #petpeeve tweets," in Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, Lisbon, Portugal: Association for Computational Linguistics, pp.2557-2563, Sep. 2015.
- E. Pavlick, P. Rastogi, J. Ganitkevitch, B. Van Durme, and C. Callison-Burch, "PPDB 2.0: Better paraphrase ranking, fine-grained entailment relations, word embeddings, and style classification," in Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing, Beijing, China: Association for Computational Linguistics, pp.425-430, Jul. 2015.
- R. Sennrich, B. Haddow, and A. Birch, "Improving neural machine translation models with monolingual data," in Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Berlin, Germany: Association for Computational Linguistics, pp. 86-96, Aug. 2016.
- M. Suzgun, L. Melas-Kyriazi, and D. Jurafsky, "Promptand-rerank: A method for zero-shot and few-shot arbitrary textual style transfer with small language models," in Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, Abu Dhabi, United Arab Emirates: Association for Computational Linguistics, pp. 2195-2222, Dec. 2022.
- A. Fabbri et al., "Improving zero and few-shot abstractive summarization with intermediate fine-tuning and data augmentation," in Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Online: Association for Computational Linguistics, pp.704-717, Jun. 2021.
- V. Gangal, S. Y. Feng, M. Alikhani, T. Mitamura, and E. Hovy, "Nareor: The narrative reordering problem," in Proceedings of the AAAI Conference on Artificial Intelligence, pp.10645-10653, 2022.
- H. Dai et al., "AugGPT: Leveraging ChatGPT for Text Data Augmentation," 2023.
- Y. Wang, Q. Yao, J. T. Kwok, and L. M. Ni, "Generalizing from a few examples: A survey on few-shot learning," ACM Computing Surveys (CSUR), Vol.53, No.3, pp.1-34, 2020. https://doi.org/10.1145/3386252
- W. Yin, J. Hay, and D. Roth, "Benchmarking zero-shot text classification: Datasets, evaluation and entailment approach," CoRR, Vol. abs/1909.00161, 2019,
- L. Cui, Y. Wu, J. Liu, S. Yang, and Y. Zhang, "Templatebased named entity recognition using BART," in Findings of the Association for Computational Linguistics: ACLIJCNLP 2021, pp.1835-1845, 2021.
- N. S. Keskar, B. McCann, L. R. Varshney, C. Xiong, and R. Socher, "Ctrl: A conditional transformer language model for controllable generation," arXiv preprint arXiv:1909. 05858, 2019.
- T. Gao, X. Yao, and D. Cehn, "SimCSE: Simple contrastive learning of sentence embeddings," 2021 Conference on Empirical Methods in Natural Language Processing, EMNLP 2021. Association for Computational Linguistics (ACL), 2021.
- C. Liu, "Learning a few-shot embedding model with contrastive learning," Proceedings of the AAAI Conference on Artificial Intelligence, Vol.35, No.10, pp.8635-8643, May 2021.
- J. Zhang et al., "Few-shot intent detection via contrastive pre-training and fine-tuning," Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pp.1906-1912, 2021.
- Park, Jangwon, KoELECTRA: Pretrained ELECTRA Model for Korean [Internet], https://github.com/monologg/Ko ELECTRA
- Ildoo Kim and Gunsoo Han and Jiyeon Ham and Woonhyuk Baek,KoGPT: KakaoBrain Korean(hangul) Generative Pre-trained Transformer [Internet], https://github.com/kakaobrain/kogpt
- Lee, Junbum, "KcBERT: Korean Comments BERT," Proceedings of the 32nd Annual Conference on Human and Cognitive Language Technology, pp.437-440, 2020