과제정보
This work was supported by Institute of Information & communications Technology Planning & Evaluation(IITP) grant funded by the Korea government(MSIT) (No. 2020-0-00368, A Neural-Symbolic Model for Knowledge Acquisition and Inference Techniques) and supported by Basic Science Research Program through the National Research Foundation of Korea(NRF) funded by the Ministry of Education(NRF-2021R1A6A1A03045425).
참고문헌
- Vaswani et al. (2017). Attention is all you need. Advances in neural information processing systems, pp. 5998-6008.
- K. Cho & M. Esipova. (2016). Can neural machine translation do simultaneous translation? arXiv preprint arXiv:1606.02012.
- Ma et al. (2018) Stacl: Simultaneous translation with implicit anticipation and controllable latency using prefix-to-prefix framework. arXiv preprint arXiv:1810.08398.
- B. Zheng, K. Liu, R. Zheng, M. Ma, H. Liu & L. Huang. (2020). Simultaneous translation policies: From fixed to adaptive. arXiv preprint arXiv:2004.13169.
- N. Arivazhagan, C. Cherry, W. Macherey & G. Foster. (2020). Retranslation versus streaming for simultaneous translation. arXiv preprint arXiv:2004.03643.
- M. Elbayad, L. Besacier & J. Verbeek. (2020). Efficient wait-k models for simultaneous machine translation. arXiv preprint arXiv:2005.08595.
- B. Zheng, R. Zheng, M. Ma & L. Huang, "Simpler and faster learning of adaptive policies for simultaneous translation," arXiv preprint arXiv:1909.01559, 2019.
- S. Li, J. Hu, B. Wang, X. Shi & Y. Chen. (2021). Xmu's simultaneous translation system at naacl 2021. Proceedings of the Second Workshop on Automatic Simultaneous Translation, pp. 19-23.
- C.-C. Chiu & C. Raffel. (2017). Monotonic chunkwise attention," arXiv preprint arXiv:1712.05382.
- Arivazhagan et al. (2019). Monotonic infinite lookback attention for simultaneous machine translation. arXiv preprint arXiv:1906.05218.
- X. Ma, J. Pino, J. Cross, L. Puzon & J. Gu. (2019). Monotonic multihead attention. arXiv preprint arXiv:1909.12406.
- H. Satija & J. Pineau. (2016). Simultaneous machine translation using deep reinforcement learning. ICML 2016 Workshop on Abstraction in Reinforcement Learning.
- J. Gu, G. Neubig, K. Cho & V. O. Li. (2016). Learning to translate in real-time with neural machine translation. arXiv preprint arXiv:1610.00388.
- B. Zheng, R. Zheng, M. Ma & L. Huang. (2019). Simultaneous translation with flexible policy via restricted imitation learning. arXiv preprint arXiv:1906.01135.
- R. J. Weiss, J. Chorowski, N. Jaitly, Y. Wu & Z. Chen. (2017). Sequence-to-sequence models can directly translate foreign speech. arXiv preprint arXiv:1703.08581.
- R. Ye, M. Wang & L. Li. (2021). End-to-end speech translation via cross-modal progressive training. arXiv preprint arXiv:2104.10380.
- C. Wang, Y. Wu, S. Liu, M. Zhou & Z. Yang. (2020). Curriculum pre-training for end-to-end speech translation. arXiv preprint arXiv:2004.10093.
- B. Zhang, I. Titov, B. Haddow & R. (2020). Sennrich. Adaptive feature selection for end-to-end speech translation. arXiv preprint arXiv:2010.08518.
- H. Nguyen, F. Bougares, N. Tomashenko, Y. Est'eve & L. Besacier. (2020). Investigating self-supervised pre-training for end-to-end speech translation. Interspeech 2020.
- Ren et al. (2020). Simulspeech: End-to-end simultaneous speech to text translation. Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pp. 3787-3796.
- X. Ma, J. Pino & P. Koehn. (2020). Simulmt to simulst: Adapting simultaneous text translation to end-to-end simultaneous speech translation. arXiv preprint arXiv:2011.02048.
- A. Karakanta, S. Papi, M. Negri & M. Turchi. (2021). Simultaneous speech translation for live subtitling: from delay to display," arXiv preprint arXiv:2107.08807.
- X. Ma, Y.Wang, M. J. Dousti, P. Koehn & J. Pino. (2021). Streaming simultaneous speech translation with augmented memory transformer. ICASSP 2021-2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 7523-7527.
- D. Liu, M. Du, X. Li, Y. Hu & L. Dai. (2021). The ustc-nelslip systems for simultaneous speech translation task at iwslt 2021. arXiv preprint arXiv:2107.00279.
- K. Papineni, S. Roukos, T. Ward & W.-J. Zhu. (2002). Bleu: a method for automatic evaluation of machine translation," Proceedings of the 40th annual meeting of the Association for Computational Linguistics, pp. 311-318.
- S. Banerjee & A. Lavie. (2005). Meteor: An automatic metric for mt evaluation with improved correlation with human judgments," Proceedings of the acl workshop on intrinsic and extrinsic evaluation measures for machine translation and/or summarization, pp. 65-72.
- C. Cherry & G. Foster. (2019). Thinking slow about latency evaluation for simultaneous machine translation. arXiv preprint arXiv:1906.00048.
- C. Raffel, M.-T. Luong, P. J. Liu, R. J. Weiss & D. Eck. (2017). Online and linear-time attention by enforcing monotonic alignments. International Conference on Machine Learning, pp. 2837-2846.
- T. Kudo & J. Richardson. (2018). Sentencepiece: A simple and language independent subword tokenizer and detokenizer for neural text processing. arXiv preprint arXiv:1808.06226.
- G. Kim. J. Son. J. Kim, H. Lee & H. Lim. (2021). Enhancing Korean Named Entity Recognition With Linguistic Tokenization Strategies. IEEE Access, 9, 151814-151823. https://doi.org/10.1109/ACCESS.2021.3126882
- C, Park. S. Eo. H. Moon & H. Lim. (2021, June). Should we find another model?: Improving Neural Machine Translation Performance with ONE-Piece Tokenization Method without Model Modification. In Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies: Industry Papers (pp. 97-104).
- C. Park. K. Kim, Y. Yang, M. Kang & H. Lim. (2021). Neural spelling correction: translating incorrect sentences to correct sentences for multimedia. Multimedia Tools and Applications, 80(26), 34591-34608. https://doi.org/10.1007/s11042-020-09148-2