DOI QR코드

DOI QR Code

Sign2Gloss2Text-based Sign Language Translation with Enhanced Spatial-temporal Information Centered on Sign Language Movement Keypoints

수어 동작 키포인트 중심의 시공간적 정보를 강화한 Sign2Gloss2Text 기반의 수어 번역

  • Kim, Minchae (Graduate School of Information, Yonsei University) ;
  • Kim, Jungeun (Dept. of Artificial Intelligence, Graduate School, Yonsei University) ;
  • Kim, Ha Young (Graduate School of Information, Yonsei University)
  • Received : 2022.09.29
  • Accepted : 2022.10.17
  • Published : 2022.10.31

Abstract

Sign language has completely different meaning depending on the direction of the hand or the change of facial expression even with the same gesture. In this respect, it is crucial to capture the spatial-temporal structure information of each movement. However, sign language translation studies based on Sign2Gloss2Text only convey comprehensive spatial-temporal information about the entire sign language movement. Consequently, detailed information (facial expression, gestures, and etc.) of each movement that is important for sign language translation is not emphasized. Accordingly, in this paper, we propose Spatial-temporal Keypoints Centered Sign2Gloss2Text Translation, named STKC-Sign2 Gloss2Text, to supplement the sequential and semantic information of keypoints which are the core of recognizing and translating sign language. STKC-Sign2Gloss2Text consists of two steps, Spatial Keypoints Embedding, which extracts 121 major keypoints from each image, and Temporal Keypoints Embedding, which emphasizes sequential information using Bi-GRU for extracted keypoints of sign language. The proposed model outperformed all Bilingual Evaluation Understudy(BLEU) scores in Development(DEV) and Testing(TEST) than Sign2Gloss2Text as the baseline, and in particular, it proved the effectiveness of the proposed methodology by achieving 23.19, an improvement of 1.87 based on TEST BLEU-4.

Keywords

Acknowledgement

This research was supported by a grant from the National Research Foundation of Korea (NRF) funded by the Korea government (MSIT; Ministry of Science and ICT) (2020R1F1A1071527).

References

  1. The Story on the National Institute of Korean Language, https://www.korean.go.kr/front/page/pageView.do?page_id=P000300&mn_id=202 (accessed September 20, 2022).
  2. M.G. Grif and A.V. Kugaevskikh, "Recognition of Deaf Gestures Based on a Bio-Inspired Neural Network," Journal of Physics: Conference Series, Vol. 1661, No. 1, p. 012038, 2020. https://doi.org/10.1088/1742-6596/1661/1/012038
  3. Korean Sign Language Dictionary, https://sldict.korean.go.kr/front/sign/signContentsView.do?current_pos_index=8&origin_no=6046&searchWay=&top_category=&category=CTE001005&detailCategory=005&searchKeyword=&page Index=1&pageJump Index= (accessed September 27, 2022).
  4. U. Bellugi and S. Fischer, "A Comparison of Sign Language and Spoken Language," Cognition, Vol. 1, No. 2, pp. 173-200, 1972. https://doi.org/10.1016/0010-0277(72)90018-2
  5. A. Voskou, K.P. Panousis, D. Kosmopoulos, D.N. Metaxas, and S. Chatzis, "Stochastic Transformer Networks with Linear Competing Units: Application to end-to-end SL Translation," Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 11946-11955, 2021.
  6. M. De Coster, K. D'Oosterlinck, M. Pizurica, P. Rabaey, S. Verlinden, M. Van Herreweghe, et al., "Frozen Pretrained Transformers for Neural Sign Language Translation," 18th Biennial Machine Translation Summit (MTSummit 2021), Association for Machine Translation in the Americas, pp. 88-97, 2021.
  7. N.C. Camgoz, S. Hadfield, O. Koller, H. Ney, and R. Bowden, "Neural Sign Language Translation," Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 7784-7793, 2018.
  8. S. Stoll, N.C. Camgoz, S. Hadfield, and R. Bowden, "Text2Sign: Towards Sign Language Production Using Neural Machine Translation and Generative Adversarial Networks," International Journal of Computer Vision, Vol. 128, No. 4, pp. 891-908, 2020. https://doi.org/10.1007/s11263-019-01281-2
  9. Z. Niu and B. Mak, "Stochastic Fine-Grained Labeling of Multi-state Sign Glosses for Continuous Sign Language Recognition," European Conference on Computer Vision, pp. 172-186, 2020.
  10. N.C. Camgoz, O. Koller, S. Hadfield, and R. Bowden, "Sign Language Transformers: Joint End-to-End Sign Language Recognition and Translation," Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 10023-10033, 2020.
  11. A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A.N. Gomez, et al., "Attention is All You Need," Advances in Neural Information P rocessing Systems, Vol. 30, 2017.
  12. S.K. Ko, C.J. Kim, H. Jung, and C. Cho, "Neural Sign Language Translation Based on Human Keypoint Estimation," Applied Science, Vol. 9, Issue 13, 2683, 2019. https://doi.org/10.3390/app9132683
  13. K. Yin and J. Read, "Better Sign Language Translation with STMC-Transformer," Proceedings of the 28th International Conference on Computational Linguistics, pp. 5975-5989, 2020.
  14. H. Sak, A. Senior, and F. Beaufays, "Long Short-Term Memory Based Recurrent Neural Network Architectures for Large Vocabulary Speech Recognition," arxiv Preprint, arXiv:1402.1128, 2014.
  15. J. Chung, C. Gulcehre, K. Cho, and Y. Bengio, "Empirical Evaluation of Gated Recurrent Neural Networks on Sequence Modeling," arXiv Preprint, arXiv:1412.3555, 2014.
  16. L. Pigou, S. Dieleman, P.J. Kindermans, and B. Schrauwen, "Sign Language Recognition Using Convolutional Neural Networks," European Conference on Computer Vision, pp. 572-578, 2014.
  17. R. Cui, H. Liu, and C. Zhang, "Recurrent Convolutional Neural Networks for Continuous Sign Language Recognition by Staged Optimization," Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 7361-7369, 2017.
  18. Korean Sign Language Maintenance Project: Korean Sign Language Dictionary (Record of Korean Sign Language), https://www.korean.go.kr/nkview/nklife/2017_2/27_0204.pdf (accessed September 29, 2022).
  19. Z. Cao, T. Simon, S.E. Wei, and Y. Sheikh, "Realtime Multi-Person 2D Pose Estimation Using Part Affinity Fields." Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 7291-7299, 2017.
  20. T. Simon, H. Joo, I. Matthews, and Y. Sheikh, "Hand Keypoint Detection in Single Images Using Multiview Bootstrapping." Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1145-1153, 2017.
  21. S.E. Wei, V. Ramakrishna, T. Kanade, and Y. Sheikh, "Convolutional Pose Machines," Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 4724-4732, 2016.
  22. A. Graves, S. Fernandez, F. Gomez, and J. Schmidhuber, "Connectionist Temporal Classification: Labelling Unsegmented Sequence Data with Recurrent Neural Networks." ProceedIngs of the 23rd International Conference on Machine Learning, pp. 369-376, 2006.
  23. H. Zhou, W. Zhou, Y. Zhou, and H. Li, "Spatial-Temporal Multi-Cue Network for Continuous Sign Language Recognition," Proceedings of the AAAI Conference on Artificial Intelligence, pp. 13009-13016, 2020.
  24. M.D. Zeiler, D. Krishnan, G.W. Taylor, and R. Fergus, "Deconvolutional Networks," 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, pp. 2528-2535, 2010.
  25. O. Chapelle and M. Wu, "Gradient Descent Optimization of Smoothed Information Retrieval Metrics," Information Retrieval, Vol. 13, No. 3, pp. 216-235, 2010. https://doi.org/10.1007/s10791-009-9110-3
  26. I. Goodfellow, Y. Bengio, and A. Courville, Deep Learning, MIT press, 2016.
  27. D.P. Kingma and J. Ba, "Adam: A Method for Stochastic Optimization," arXiv Preprint, arXiv: 1412.6980, 2014.
  28. K. Papineni, S. Roukos, T. Ward, and W.J. Zhu, "Bleu: A Method for Automatic Evaluation of Machine Translation," Proceedings of the 40th Annual Meeting of the Association for Computational Linguistics, pp. 311-318, 2002.
  29. C. Szegedy, V. Vanhoucke, S. Ioffe, J. Shlens, and Z. Wojna, "Rethinking the Inception Architecture for Computer Vision." Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2818-2826, 2016.
  30. R. Muller, S. Kornblith, and G.E. Hinton, "When Does Label Smoothing Help?" Advances in Neural Information P rocessing Systems, Vol. 32, 2019.
  31. B. Subramanian, B. Olimov, and J. Kim, "Fast Convergence GRU Model for Sign Language Recognition," Journal of Korea Multimedia Society, Vol. 25, No 9, pp. 1257-1265, 2022. https://doi.org/10.9717/KMMS.2022.25.9.1257