Acknowledgement
본 논문은 2022년도 강릉원주대학교 신임교원 연구비 지원과 2023년도 교육부의 재원으로 한국연구재단의 지원을 받아 수행된 지자체-대학 협력 기반 지역 혁신 사업의 결과입니다(2022RIS-005).
References
- L. Delphin-Poulat and C. Plapous, "Mean teacher with data augmentation for dcase 2019 task 4," Orange Labs Lannion, Tech. Rep., 2019.
- A. Tarvainen and H. Valpola, "Mean teachers are better role models: Weight-averaged consistency targets improve semi-supervised deep learning results," Proc. NIPS, 1-10 (2017).
- K. He, X. Zhang, S. Ren, and J. Sun, "Deep residual learning for image recognition," Proc. IEEE Conf. on CVPR, 770-778 (2016).
- K. He, X. Zhang, S. Ren, and J. Sun, "Identity mappings in deep residual networks," Proc. Computer Vision-ECCV, 1-15 (2016).
- S. Zagoruyko and N. Komodakis, "Wide residual networks," arXiv preprint arXiv:1605.07146 (2016).
- G. Huang, Z. Liu, L. Van Der Maaten, and K. Q. Weinberger, "Densely connected convolutional networks," Proc. Computer Vision-ECCV, 1-9 (2017).
- B. McMahan and D. Rao, "Listening to the world improves speech command recognition," Proc. AAAI Conf. on Artificial Intelligence, 378-385 (2018).
- K. Palanisamy, D. Singhania, and A. Yao, "Rethinking CNN models for audio classification," arXiv preprint arXiv:2007.11154 (2020).
- PyTorch Torch.nn.GRU, https://pytorch.org/docs/stable/generated/torch.nn.GRU.html, (Last viewed February 12, 2023).
- A PyTorch Implementation for Densely Connected Convolutional Networks (DenseNets), https://github.com/andreasveit/densenet-pytorch, (Last viewed February 12, 2023).
- S. Ioffe and C. Szegedy, "Batch normalization: Accelerating deep network training by reducing internal covariate shift," arXiv preprint arXiv:1502.03167 (2015).
- X. Glorot, A. Bordes, and Y. Bengio, "Deep sparse rectifier neural networks," Proc. of the 14th International Conf. on Artificial Intelligence and Statistics, 315-323 (2011).
- N. Srivastava, G. Hinton, A. Krizhevsky, I. Sutskever, and R. Salakhutdinov, "Dropout: a simple way to prevent neural networks from overfitting," J. Mach. Learn. Res. 15, (2014).
- DCASE 2020 Task 4 GitHub, https://github.com/turpaultn/dcase20_task4, (Last viewed February 12, 2023).
- N. Turpault, R. Serizel, A. Shah, and J. Salamon, "Sound event detection in domestic environments with weakly labeled data and soundscape synthesis," Proc. DCASE Workshop, 253-257 (2019).
- S. Park and M. Elhilali, "Time-balanced focal loss for audio event detection," Proc. ICASSP, 311-315 (2022).
- X. Glorot and Y. Bengio, "Understanding the difficulty of training deep feedforward neural networks," Proc. 13th International Conf. on Artificial Intelligence and Statistics, 249-256 (2010).
- K. He, X. Zhang, S. Ren, and J. Sun, "Delving deep into rectifiers: Surpassing human-level performance on imagenet classification," Proc. IEEE ICCV, 1026-1032 (2015).
- DCASE 2020 Task 4: Sound Event Detection and Separation in Domestic Environments, https://dcase.community/challenge2020/task-sound-event-detection-and-separation-in-domestic-environments, (Last viewed July 25, 2023).
- A. Mesaros, T. Heittola, T. Virtanen, and M. D. Plumbley, "Sound event detection: A tutorial," IEEE Signal Process. Mag. 38, 67-83 (2021). https://doi.org/10.1109/MSP.2021.3090678
- D. P. Kingma and J. Ba, "Adam: A method for stochastic optimization," arXiv preprint arXiv:1412.6980 (2014).