Browse > Article
http://dx.doi.org/10.5909/JBE.2020.26.7.868

An Analysis on the Properties of Features against Various Distortions in Deep Neural Networks  

Kang, Jung Heum (Department of Computer Science and Engineering, Kyung Hee Univ.)
Jeong, Hye Won (Department of Computer Science and Engineering, Kyung Hee Univ.)
Choi, Chang Kyun (Department of Computer Science and Engineering, Kyung Hee Univ.)
Ali, Muhammad Salman (Department of Computer Science and Engineering, Kyung Hee Univ.)
Bae, Sung-Ho (Department of Computer Science and Engineering, Kyung Hee Univ.)
Kim, Hui Yong (Department of Computer Science and Engineering, Kyung Hee Univ.)
Publication Information
Journal of Broadcast Engineering / v.26, no.7, 2021 , pp. 868-876 More about this Journal
Abstract
Deploying deep neural network model training performs remarkable performance in the fields of Object detection and Instance segmentation. To train these models, features are first extracted from the input image using a backbone network. The extracted features can be reused by various tasks. Research has been actively conducted to serve various tasks by using these learned features. In this process, standardization discussions about encoding, decoding, and transmission methods are proceeding actively. In this scenario, it is necessary to analyze the response characteristics of features against various distortions that may occur in the data transmission or data compression process. In this paper, experiment was conducted to inject various distortions into the feature in the object recognition task. And analyze the mAP (mean Average Precision) metric between the predicted value output from the neural network and the target value as the intensity of various distortions was increased. Experiments have shown that features are more robust to distortion than images. And this points out that using the feature as transmission means can prevent the loss of information against the various distortions during data transmission and compression process.
Keywords
Convolutional neural networks; Feature compression; Data transmission; Image restoration; Instance segmentation;
Citations & Related Records
연도 인용수 순위
  • Reference
1 Ren, S., He, K., Girshick, R., & Sun, J. ''Faster r-cnn: Towards real-time object detection with region proposal networks.'' Advances in neural information processing systems, 28, pp.91-99, 2015
2 Duan, L., Liu, J., Yang, W., Huang, T., & Gao, W. "Video coding for machines: A paradigm of collaborative compression and intelligent analytics.'' IEEE Transactions on Image Processing, 29, 8680-8695, 2020.   DOI
3 Girshick, R., Donahue, J., Darrell, T., & Malik, J. " Rich feature hierarchies for accurate object detection and semantic segmentation.'' In Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 580-587, 2014.
4 A. G. Howard, M. Zhu, B. Chen, D. Kalenichenko, W. Wang, T. Weyand, M. Andreetto, and H. Adam, ''MobileNets: Efficient convolutional neural networks for mobile vision applications,'' 2017, arXiv:1704.04861.
5 Lin, T. Y., Maire, M., Belongie, S., Hays, J., Perona, P., Ramanan, D., ... & Zitnick, C. L. "Microsoft coco: Common objects in context.'' In European conference on computer vision, pp. 740-755, 2014.
6 Fan, L., Zhang, F., Fan, H., & Zhang, C. "Brief review of image denoising techniques.'' Visual Computing for Industry, Biomedicine, and Art, 2(1), 1-12, 2019.   DOI
7 Kuznetsova, Alina, et al. "The open images dataset v4." International Journal of Computer Vision 128.7, 1956-1981, 2020.   DOI
8 Lin, T. Y., Goyal, P., Girshick, R., He, K., & Dollar, P. " Focal loss for dense object detection. '' In Proceedings of the IEEE international conference on computer vision. p. 2980-2988, 2017.
9 He, K., Gkioxari, G., Dollar, P., & Girshick, R. "Mask r-cnn.'' In Proceedings of the IEEE international conference on computer vision, pp. 2961-2969, 2017.
10 He, K., Zhang, X., Ren, S., & Sun, J. "Deep residual learning for image recognition.'' In Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 770-778, 2016.
11 Girshick, R. "Fast r-cnn.'' In Proceedings of the IEEE international conference on computer vision, pp. 1440-1448, 2015.
12 Yang, W., Zhang, X., Tian, Y., Wang, W., Xue, J. H., & Liao, Q. "Deep learning for single image super-resolution: A brief review. '' IEEE Transactions on Multimedia, 21(12), 3106-3121, 2019.   DOI
13 Lainema, J., Bossen, F., Han, W. J., Min, J., & Ugur, K. "Intra coding of the HEVC standard.'' IEEE transactions on circuits and systems for video technology, 22(12), 1792-1801, 2012.   DOI
14 Lin, T. Y., Dollar, P., Girshick, R., He, K., Hariharan, B., & Belongie, S. "Feature pyramid networks for object detection.'' In Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 2117-2125, 2017.
15 Y. Kang, J. Hauswald, C. Gao, A. Rovinski, T. Mudge, J. Mars,and L. Tang, "Neurosurgeon: Collaborative intelligence between the cloud and mobile edge." in Proc. 22nd ACM Int.Conf. Arch. Support Programming Languages and Operating System, pp. 615-629, 2017
16 Bajic, Ivan V., Weisi Lin, and Yonghong Tian. "Collaborative intelligence: Challenges and opportunities.'' ICASSP 2021-2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, 2021.
17 Jacob, B., Kligys, S., Chen, B., Zhu, M., Tang, M., Howard, A.,et.al, "Quantization and training of neural networks for efficient integer-arithmetic-only inference.'' In Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 2704-2713, 2018.