과제정보
이 논문은 2023년도 정부(산업통상자원부)의 재원으로 한국에너지기술평가원의 지원을 받아 수행된 연구임(20224B10100060, 회전설비 인공지능형 진동 감시 시스템 개발)
참고문헌
- Bum-Jun kim et al. (2018). Research on data augmentation method for audio marking based on deep neural network. Journal of the Korean Acoustical Society, 37(6), 475-482.
- Ju-Hyuk Han et al. (2022). MR-CL: Contrastive Loss-based multimodal convergence learning method for emotion recognition. Korean Society of Information Scientists and Engineers Academic Presentation Papers
- Chung-bin Kim et al. (2022). Feature extraction from self-supervised learning model and transformer-based multimodal expression learning for combined voice-text emotion recognition. Korean Society of Information Scientists and Engineers Academic Presentation Papers
- Yong-hwa Jo el al. (2020). Implementation of a Classification System for Dog Behaviors using YOLI-based Object Detection and a Node.js Server. Journal of Convergence Signal Processing Society. v.21, no.1, pp.29-37
- Lee-sun Mun et al. (2023). Multimodal emotion recognition based on biometric signals and voice data. Proceedings of the Domestic Conference of the Society of Control and Robotics Systems
- P. Tzirakis, G. Trigeorgis, M. A. Nicolaou, B. W. Schuller and S. Zafeiriou, "End-to-End Multimodal Emotion Recognition Using Deep Neural Networks," in IEEE Journal of Selected Topics in Signal Processing, vol. 11, no. 8, pp. 1301-1309, Dec. 2017, https://doi.org/10.1109/JSTSP.2017.2764438
- H. Ranganathan, S. Chakraborty and S. Panchanathan, "Multimodal emotion recognition using deep learning architectures," 2016 IEEE Winter Conference on Applications of Computer Vision (WACV), Lake Placid, NY, USA, 2016, pp. 1-9,
- Kahou, S.E., Bouthillier, X., Lamblin, P. et al. EmoNets: Multimodal deep learning approaches for emotion recognition in video. J Multimodal User Interfaces 10, 99-111 (2016). https://doi.org/10.1007/s12193-015-0195-2
- Xu, Haiyang, et al. "Learning alignment for multimodal emotion recognition from speech."arXiv preprint arXiv:1909.05645 (2019).
- Alu, D. A. S. C., Elteto Zoltan, and Ioan Cristian Stoica. "Voice based emotion recognition with convolutional neural networks for companion robots." Science and Technology 20.3 (2017): 222-240.
- Livingstone, Steven R., and Frank A. Russo. "The Ryerson Audio-Visual Database of Emotional Speech and Song (RAVDESS): A dynamic, multimodal set of facial and vocal expressions in North American English." PloS one 13.5 (2018): e0196391.
- Feichtenhofer, Christoph, et al. "Slowfast networks for video recognition." Proceedings of the IEEE/CVF international conference on computer vision. 2019.
- Shen, Jonathan, et al. "Natural tts synthesis by conditioning wavenet on mel spectrogram predictions." 2018 IEEE international conference on acoustics, speech and signal processing (ICASSP). IEEE, 2018.
- Schluter, Jan, and Thomas Grill. "Exploring data augmentation for improved singing voice detection with neural networks." ISMIR. 2015.
- Ngiam, Jiquan, et al. "Multimodal deep learning." Proceedings of the 28th international conference on machine learning (ICML-11). 2011.
- Jae-Eun Lee et al. (2022) A neck healthy warning algorithm for identifying text neck posture prevention. Journal of Convergence Signal Processing Society. v.23, no.3, pp.115-122
- HATAMI, Nima; GAVET, Yann; DEBAYLE, Johan. Classification of time-series images using deep convolutional neural networks. In: Tenth international conference on machine vision (ICMV 2017). SPIE, 2018. p. 242-249.