Browse > Article
http://dx.doi.org/10.6109/jkiice.2019.23.1.55

Conv-XP Pruning of CNN Suitable for Accelerator  

Woo, Yonggeun (School of Computer Science and Engineering, Korea University of Technology and Education)
Kang, Hyeong-Ju (School of Computer Science and Engineering, Korea University of Technology and Education)
Abstract
Convolutional neural networks (CNNs) show high performance in the computer vision, but they require an enormous amount of operations, making them unsuitable for some resource- or energy-starving environments like the embedded environments. To overcome this problem, there have been much research on accelerators or pruning of CNNs. The previous pruning schemes have not considered the architecture of CNN accelerators, so the accelerators for the pruned CNNs have some inefficiency. This paper proposes a new pruning scheme, Conv-XP, which considers the architecture of CNN accelerators. In Conv-XP, the pruning is performed following the 'X' or '+' shape. The Conv-XP scheme induces a simple architecture of the CNN accelerators. The experimental results show that the Conv-XP scheme does not degrade the accuracy of CNNs, and that the accelerator area can be reduced by 12.8%.
Keywords
Neural networks; Convolutional neural networks; Accelerator; Pruning;
Citations & Related Records
Times Cited By KSCI : 1  (Citation Analysis)
연도 인용수 순위
1 D. Kim, J. Ahn, and S. Yoo, "A novel zero weight/ activation-aware hardware architecture of convolutional neural network," in Proceedings of Design, Automation & Test in Europe Conference & Exhibition, pp. 1462-1467, 2017.
2 V. Lebedev and V. Lempitsky, "Fast ConvNets using group-wise brain damage," in Proceedings of Computer Vision and Pattern Recognition, pp. 2554-2564, 2016.
3 S. Anwar, K. Hwang, and W. Sung, "Structured pruning of deep convolutional neural networks," ACM Journal on Emerging Technologies in Computing System, vol. 13, no. 3, pp. 32:1-32:18, Feb. 2017.
4 Y. -J. Kim and E. -G. Kim, "Image based Fire Detection using Convolutional Neural Network," Journal of the Korea Institute of Information and Communication Engineering, vol. 20, no. 9, pp. 1649-1656, Sep. 2016.   DOI
5 A. Krizhevsky, I. Sutskever, and G. E. Hinton, "ImageNet classification with deep convolutional neural networks," in Proceedings of Advances in Neural Information Processing Systems, pp. 1097-1105, 2012.
6 K. Simonyan and A. Zisserman, "Very deep convolutional networks for large-scale image recognition," in Proceedings of International Conference on Learning Representations, pp. 1-14, 2015.
7 C. Szegedy, W. Liu, Y. Jia, P. Sermanet, S. Reed, D. Anguelov, D. Erhan, V. Vanhoucke, and A. Rabinovich, "Going deeper with convolutions," in Proceedings of Computer Vision and Pattern Recognition, pp. 1-9, 2015.
8 K. He, X. Zhang, S. Ren, and J. Sun, "Deep residual learning for image recognition," in Proceedings of Computer Vision and Pattern Recognition, pp. 770-778, 2016.
9 S. Han, J. Pool, J. Tran, and W. J. Dally, "Learning both weights and connections for efficient neural networks," in Proceedings of Advances in Neural Information Processing Systems, pp. 1135-1143, 2015.
10 S. Han, H. Mao, and W. J. Dally, "Deep Compression: Compressing deep neural networks with pruning, trained quantization and Huffman coding," in Proceedings of International Conference on Learning Representations, pp. 1-14, 2016.
11 W. Wen, C. Wu, Y. Wang, Y. Chen, and H. Li, "Learning structured sparsity in deep neural networks," in Proceedings of Advances in Neural Information Processing Systems, pp. 2074-2082, 2016.
12 Y. He, X. Zhang, and J. Sun, "Channel pruning for accelerating very deep neural networks," in Proceedings of International Conference on Computer Vision, pp. 1398-1406, 2017.
13 S. -H. Kwon, K. -W. Park, and B. -H. Chang, "A Comparison of Predicting Movie Success between Artificial Neural Network and Decision Tree," Asia-pacific Journal of Multimedia Services Convergent with Art, Humanities, and Sociology, vol. 7, no. 4, pp. 593-601, Apr. 2017.
14 N. Yu, S. Qiu, X. Hu, and J. Li, "Accelerating convolutional neural networks by group-wise 2D-filter pruning," in Proceedings of International Joint Conference on Neural Networks, pp. 2502-2509, 2017.
15 H. Li, A. Kadav, I. Durdanovic, H. Samet, and H. P. Graf, "Pruning filters for efficient ConvNets," in Proceedings of International Conference on Learning Representations, pp. 1-13, 2017.
16 J. Yu, A. Lukefahr, D. Palframan, G. Dasika, R. Das, and S. Mahlke, "Scalpel: Customizing DNN pruning to the underlying hardware parallelism," in Proceedings of International Symposium on Computer Architecture, pp. 548-560, 2017.
17 P. Molchanov, S. Tyree, T. Karras, T. Aila, and J. Kautz, "Pruning convolutional neural networks for resource efficient inference," in Proceedings of International Conference on Learning Representations, pp. 1-17, 2017.
18 S. Liu, Z. Du, J. Tao, D. Han, T. Luo, Y. Zie, Y. Chen, and T. Chen, "Cambricon: An instruction set architecture for neural networks," in Proceedings of International Symposium on Computer Architecture, pp. 393-405, 2016.
19 T. Chen, Z. Du, N. Sun, J. Wang, C. Wu, Y. Chen, and O. Temam, "DianNao: A small-footprint high-throughput accelerator for ubiquitous machine-learning," in Proceedings of International Conference on Architectural Support for Programming Languages and Operating Systems, pp. 269-283, 2014.
20 Y. Chen, T. Luo, S. Liu, S. Zhang, L. He, J. Wang, L. Li, T. Chen, Z. Xu, N. Sun, and O. Temam, "DaDianNao: A machine-learning supercomputer," in Proceedings of International Symposium on Micro-architecture, pp. 609-622, 2014.
21 C. Zhang, P. Li, G. Sun, Y. Guan, B. Xiao, and J. Cong, "Optimizing FPGA-based accelerator design for deep convolutional neural networks," in Proceedings of ACM/SIGDA International Symposium on Field Programmable Gate Arrays, pp. 161-170, 2015.
22 S. Han, X. Liu, H. Mao, J. Pu, A. Pedram, M. A. Horowitz, and W. J. Dally, "EIE: Efficient inference engine on compressed deep neural network," in Proceedings of International Symposium on Computer Architecture, pp. 243-254, 2016.
23 J. Albericio, P. Judd, T. Hetherington, T. Aamodt, N. E. Jerger, and A. Moshovos, "Cnvlutin: Ineffectual-neuron-free deep neural network computing," in Proceedings of International Symposium on Computer Architecture, pp. 1-13, 2016.
24 L. Song, Y. Wang, Y. Han, X. Zhao, B. Liu, and X. Li, "C-Brain: A deep learning accelerator that tames the diversity of CNNs through adaptive data-level parallelization," in Proceedings of Design Automation Conference, pp. 123:1-123:6, 2016.
25 J. Qiu, J. Wang, S. Yao, K. Guo, B. Li, E. Zhou, J. Yu, T. Tang, N. Xu, S. Song, Y. Wang, and H. Yang, "Going deeper with embedded FPGA platform for convolutional neural network," in Proceedings of ACM/SIGDA International Symposium on Field Programmable Gate Arrays, pp. 26-35, 2016.
26 M. Motamedi, P. Gysel, V. Akella, and S. Ghiasi, "Design space exploration of FPGA-based deep convolutional neural networks," in Proceedings of Asia and South Pacific Design Automation Conference, pp. 575-580, 2016.
27 J. Jo, S. Cha, D. Rho, and I.-C. Park, "DSIP: A scalable inference accelerator for convolutional neural networks," IEEE Journal of Solid-State Circuits, vol. 53, no. 2, pp. 605-618, Feb. 2018.   DOI
28 S. Han, J. Kang, H. Mao, Y. Hu, X. Li, Y. Li, D. Xie, H. Luo, S. Yao, Y. Wang, H. Yang, and W. J. Dally, "ESE: Efficient speech recognition engine with sparse LSTM on FPGA," in Proceedings of ACM/SIGDA International Symposium on Field Programmable Gate Arrays, pp. 75-84, 2017.
29 S. Zhang, Z. Du, L. Zhang, H. Lan, S. Liu, L. Li, Q. Guo, T. Chen, and Y. Chyen, "Cambricon-X: An accelerator for sparse neural networks," in Proceedings of International Symposium on Micro-architecture, pp. 20:1-20:12, 2016.