Acknowledgement
이 논문은 2021년도 중앙대학교 CAU GRS 지원에 의하여 작성되었음. 이 성과는 과학기술정보통신부의 재원으로 한국연구재단의 지원을 받아 수행된 연구임 (NRF-2021R1F1A1056516).
References
- Malik M, Malik MK, Mehmood K, and Makhdoom I (2020). Automatic speech recognition: A survey, Multimedia Tools and Applications, 80, 9411-9457. https://doi.org/10.1007/s11042-020-10073-7
- Kim YG and Lee JW (2020). Development of Korean automatic speech recognition model using transformer, Proceedings of Symposium of the Korean Institute of Communications and Information Sciences, 71, 659-660.
- Lee SG and Kwon SI (2014). Elderly speech analysis for improving elderly speech recognition, Communications of the Korean Institute of Information Scientists and Engineers, 32, 16-20.
- Young V and Mihailidis A (2010). Difficulties in automatic speech recognition of dysarthric speakers and implications for speech-based applications used by the elderly: A literature review, Assistive Technology, 22, 99-112. https://doi.org/10.1080/10400435.2010.483646
- Audacity Team, Audacity® (2020). Free audio editor and recorder [Computer application] Version 2.4.1. 2020.
- Chi YK, Han JW, Jeong HJ et al. (2014). Development of a screening algorithm for Alzheimer's disease using categorical verbal fluency, PLOS ONE, 9, e84111.
- Chakraborty K, Talele A, and Upadhya S (2014). Voice recognition using MFCC algorithm, International Journal of Innovative Research in Advanced Engineering (IJIRAE), 1, 2349-2163.
- Chan W, Jaitly N, Quoc VL, and Vinyals O (2016). Listen, attend and spell: A neural network for large vocabulary conversational speech recognition, In proceedings of 2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Shanghai, 4960-4964.
- Dong L, Xu S, and Xu B (2018). Speech-transformer: A no-recurrence sequence-to-sequence model for speech recognition, In proceedings of 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Calgary, AB, 5884-5888.
- Han W, Zhang Z, Zhang Y et al. (2020). Contextnet: Improving convolutional neural networks for automatic speech recognition with global context, Available from: arXiv preprint arXiv:2005.03191..
- Wu Z, Liu Z, Lin J, Lin Y, and Han S (2020). Lite transformer with long-short range attention, Available from: arXiv preprint arXiv:2004.11886, 2020.
- Lu Y, Li Z, He D, Sun Z, Dong B, Qin T, Wang L, and Liu TY (2019). Understanding and improving transformer from a multi-particle dynamic system point of view, Available from: arXiv preprint arXiv:1906.02762.
- Gulati A, Qin J, Chiu CC et al. (2020). Conformer: Convolution-augmented transformer for speech recognition, Available from: arXiv preprint arXiv:2005.08100.
- Graves A and Jaitly N (2013). Towards end-to-end speech recognition with recurrent neural networks, Proceedings of the 31st International Conference on International Conference on Machine Learning PMLR, 32, 1764-1772.
- Kim KW, Ji YG, and Han JW (2014). inventors. Method of diagnosing dementia based on verbal fluency and apparatus therefore. South Korea patent KR101437569B1, 2014 Sep 04.
- Panayotov V, Chen G, Povey D, and Khudanpur S (2015). Librispeech: An ASR corpus based on public domain audio books, In proceedings of 2015 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Brisbane, 5206-5210.
- Park DS, Chan W, Zhang Y, Chiu C, Zoph B, Cubuk ED, and Quoc VL (2019). Specaugment: A simple data augmentation method for automatic speech recognition, Available from: arXiv preprint arXiv:1904.08779