Acknowledgement
This work was supported by Korea Research Institute for defense Technology planning and advancedment (KRIT) grant funded by the Korea government (DAPA (Defense Acquisition Program Administration)) (No. 21-107-E00-009-02, "Realtime complex battlefield situation awareness technology")
References
- C.-E. Lee, J. Baek, J. Son, and Y.-G. Ha, " Deep AI military staff: Cooperative battlefield situation awareness for commander's decision making," The Journal of Supercomputing, vol. 79, no. 6, pp. 6040-6069, Apr., 2023, DOI: 10.1007/s11227-022-04882-w.
- A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. N. Gomez, L. Kaiser, and I. Polosukhin, "Attention is all you need," Advances in neural information processing systems, vol. 30, 2017, [Online], https://proceedings.neurips.cc/paper_files/paper/2017/file/3f5ee243547dee91fbd053c1c4a845aa-Paper.pdf.
- Y. Li, "Deep reinforcement learning: An overview," arXiv, 2017, [Online], https://arxiv.org/abs/1701.07274.
- DARPA, "Defense Advanced Research Projects Agency," [Online], https://www.darpa.mil, Accessed: 09 23, 2023.
- Association of Research Libraries, "Establish a Universal, Open Library or Digital Data Commons," [Online], https://www.arl.org/resources/establish-a-universal-open-library-or-digital-data-commons/, Accessed: 09 23, 2023.
- W. Guo, J. Wang, and S. Wang, "Deep Multimodal Representation Learning: A Survey," IEEE Access, vol. 7, pp. 63373-63394, May, 2019, DOI: 10.1109/ACCESS.2019.2916887.
- R. Krishna, Y. Zhu, O. Groth, J. Johnson, K. Hata, J. Kravitz, S. Chen, Y. Kalantidis, L.-J. Li, D. A. Shamm, M. S. Bernstein, and L. Fei-Fei, "Visual genome: Connecting language and vision using crowdsourced dense image annotations," International journal of computer vision, vol. 123, pp. 32-73, Feb., 2017, DOI: 10.1007/s11263-016-0981-7.
- Z. Wang and J.-C. Liu, "Translating math formula images to latex sequences using deep neural networks with sequence-level training," International Journal on Document Analysis and Recognition (IJDAR), vol. 24, no. 1-2, pp. 63-75, Noc., 2021, DOI: 10.1007/s10032-020-00360-2.
- J. Devlin, M.-W. Chang, K. Lee, and K. Toutanova, "BERT: Pre-training of deep bidirectional transformers for language understanding," arXiv:1810.04805, 2018, [Online], https://arxiv.org/abs/1810.04805.