DOI QR코드

DOI QR Code

Sentence Recommendation Using Beam Search in a Military Intelligent Image Analysis System

군사용 지능형 영상 판독 시스템에서의 빔서치를 활용한 문장 추천

  • Received : 2021.10.07
  • Accepted : 2021.10.23
  • Published : 2021.11.30

Abstract

Existing image analysis systems in use in the military field are carried out by readers analyzing and identifying images themselves, writing and disseminating related content, and in this process, repetitive tasks are frequent, resulting in workload. In this paper, to solve the previous problem, we proposed an algorithm that can operate the Seq2Seq model on a word basis, which operates on a sentence basis, and applied the Attention technique to improve accuracy. In addition, by applying the Beam Search technique, we would like to recommend various current identification sentences based on the past identification contents of a specific area. It was confirmed through experiments that the Beam Search technique recommends sentences more effectively than the existing greedy Search technique, and confirmed that the accuracy of recommendation increases when the size of Beam is large.

군사 분야에서 사용 중인 기존 영상 판독 시스템은 판독관들이 직접 영상을 분석 및 식별하여 관련 내용을 보고서에 작성하고 전파하는 방식으로 진행되는데 이 과정에서 반복 작업이 빈번하여 업무 과부하가 발생한다. 본 논문에서는 이러한 문제를 해결하고자, 기존의 문장 단위로 동작하는 Seq2Seq 모델을 단어 단위로 동작할 수 있는 알고리즘을 제안하고, Attention 기법을 적용해 정확도를 향상시키고자 한다. 또한 Beam 탐색 기법을 응용하여 특정 지역의 과거 식별내용을 바탕으로 현재 식별 문장을 다양하게 추천하고자 한다. 실험을 통해 Beam 탐색 기법이 기존 Greedy 탐색 기법보다 효과적으로 문장을 추천하는 것을 확인하였고, Beam의 크기가 클 때 추천의 정확도가 높아지는 것을 확인하였다.

Keywords

Acknowledgement

이 논문은 국방과학연구소의 지원을 받아 수행된 연구임(UD190025FD).

References

  1. T. H. Jeon, H. S. Na, J. H. Ahn, and D. H. Im, "Pre-processing and implementation for intelligent imagery interpretation system," Proceedings of the Korea Information Processing Society Conference, Vol.28, pp.305-307, 2021.
  2. Y. D. Kim and H. J. Gwon, "A study on defense command and control system AI application," Korea Information Processing Society Review, Vol.24, No.1, pp.13-18, 2017.
  3. I. Sutskever, O. Vinyals, and V. L. Quoc, "Sequence to sequence learning with neural networks," In: Advances in neural Information Processing Systems, pp.3104-3112, 2014.
  4. M. Zhang, Z. Li, G. Fu, and M. Zhang, "Syntax-enhanced neural machine translation with syntax-aware word representations," Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Vol.1 (Long and Short Papers), 2019.
  5. K. Palasundram, N. M. Sharef, N. Nasharuddin, K. Kasmiran, and A. Azman "Sequence to sequence model performance for education chatbot," International Journal of Emerging Technologies in Learning (iJET), Vol.14, No.24, pp.56-68, 2019.
  6. K. Qian and Z. Yu. "Domain adaptive dialog generation via meta learning," Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, 2019.
  7. D. Bahdanau, C. Kyunghyun, and Y. Bengio, "Neural machine translation by jointly learning to align and translate," 3rd International Conference on Learning Representations, ICLR 2015.
  8. S. Wiseman and A. M. Rush. "Sequence -to-sequence learning as beam-search optimization," Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing, 2016.
  9. pyhwp Documentation 2013. [Internet], https://pythonhosted.org/pyhwp/ko/ (accessed August 2, 2021.)
  10. P. Bojanowski, E. Grave, A. Joulin, and T. Mikolov, "Enriching word vectors with subword information," Transactions of the Association for Computational Linguistics, Vol.5, pp.135-146, 2017. https://doi.org/10.1162/tacl_a_00051
  11. E. J. Park and S. Z. Cho, "KoNLPy: Korean natural language processing in Python," Annual Conference on Human and Language Technology, pp.133-136, 2014.
  12. Comparison of Korean stemming analyzer performance (2018). [Internet], https://iostream.tistory.com/144 (accessed August 2, 2021)
  13. T. Mikolov, K. Chen, G Corrado, and J. Dean, "Efficient estimation of word representations in vector space," arXiv preprint arXiv:1301.3781, 2013.
  14. J. Pennington, R. Socher, and C. D. Manning, "Glove: Global vectors for word representation," Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing, pp.1532-1543, 2014.
  15. Y. Bengio, R. Ducharme, P. Vincent, and C. Jauvin, "A neural probabilistic language model," Journal of Machine Learning Research, Vol.3, 1137-1155, 2003.
  16. T. Mikolov, M. Karafiat, L. Burget, J. Cernocky, and S. Khudanpur, "Recurrent neural network based language model," In Eleventh Annual Conference of the International Speech Communication Association, Vol.9, pp.1045-1048, 2010.
  17. Y. Bengio, P. Simard, and P. Frasconi, "Learning long-term dependencies with gradient descent is difficult." IEEE Transactions on Neural Networks, Vol.5, No.2, pp.157-166, 1994. https://doi.org/10.1109/72.279181
  18. S. Hochreiter and J. Schmidhuber, "Long short-term memory," Neural Computation, Vol.9, No.8, pp.1735-1780, 1997. https://doi.org/10.1162/neco.1997.9.8.1735
  19. J. Chung, C. Gulcehre, K. Cho, and Y. Bengio, "Empirical evaluation of gated recurrent neural networks on sequence modeling." NIPS 2014 Workshop on Deep Learning, Dec. 2014.
  20. S. Mangal, P. Joshi, and R. Modak, "Lstm vs. gru vs. bidirectional rnn for script generation," arXiv preprint arXiv: 1908.04332, 2019.
  21. K. Cho, "Noisy parallel approximate decoding for conditional recurrent language model," arXiv preprint arXiv:1605.03835, 2016.