Fig. 1 Operation architecture of the accelerator in [21]
Fig. 2 Multiplier connection of a sparse accelerator for the conventionally pruned networks
Fig. 3 Conv-XP pruning pattern: (a) ‘X’ pattern and (b)‘+’ pattern
Fig. 4 Multiplier connection of a sparse accelerator for the networks pruned by Conv-XP
Table. 1 Classification accuracy comparison on VGG16 and ResNet-50
Table. 2 Area comparison (㎛2)
References
- Y. -J. Kim and E. -G. Kim, "Image based Fire Detection using Convolutional Neural Network," Journal of the Korea Institute of Information and Communication Engineering, vol. 20, no. 9, pp. 1649-1656, Sep. 2016. https://doi.org/10.6109/jkiice.2016.20.9.1649
- S. -H. Kwon, K. -W. Park, and B. -H. Chang, "A Comparison of Predicting Movie Success between Artificial Neural Network and Decision Tree," Asia-pacific Journal of Multimedia Services Convergent with Art, Humanities, and Sociology, vol. 7, no. 4, pp. 593-601, Apr. 2017.
- A. Krizhevsky, I. Sutskever, and G. E. Hinton, "ImageNet classification with deep convolutional neural networks," in Proceedings of Advances in Neural Information Processing Systems, pp. 1097-1105, 2012.
- K. Simonyan and A. Zisserman, "Very deep convolutional networks for large-scale image recognition," in Proceedings of International Conference on Learning Representations, pp. 1-14, 2015.
- C. Szegedy, W. Liu, Y. Jia, P. Sermanet, S. Reed, D. Anguelov, D. Erhan, V. Vanhoucke, and A. Rabinovich, "Going deeper with convolutions," in Proceedings of Computer Vision and Pattern Recognition, pp. 1-9, 2015.
- K. He, X. Zhang, S. Ren, and J. Sun, "Deep residual learning for image recognition," in Proceedings of Computer Vision and Pattern Recognition, pp. 770-778, 2016.
- S. Han, J. Pool, J. Tran, and W. J. Dally, "Learning both weights and connections for efficient neural networks," in Proceedings of Advances in Neural Information Processing Systems, pp. 1135-1143, 2015.
- S. Han, H. Mao, and W. J. Dally, "Deep Compression: Compressing deep neural networks with pruning, trained quantization and Huffman coding," in Proceedings of International Conference on Learning Representations, pp. 1-14, 2016.
- W. Wen, C. Wu, Y. Wang, Y. Chen, and H. Li, "Learning structured sparsity in deep neural networks," in Proceedings of Advances in Neural Information Processing Systems, pp. 2074-2082, 2016.
- N. Yu, S. Qiu, X. Hu, and J. Li, "Accelerating convolutional neural networks by group-wise 2D-filter pruning," in Proceedings of International Joint Conference on Neural Networks, pp. 2502-2509, 2017.
- H. Li, A. Kadav, I. Durdanovic, H. Samet, and H. P. Graf, "Pruning filters for efficient ConvNets," in Proceedings of International Conference on Learning Representations, pp. 1-13, 2017.
- Y. He, X. Zhang, and J. Sun, "Channel pruning for accelerating very deep neural networks," in Proceedings of International Conference on Computer Vision, pp. 1398-1406, 2017.
- J. Yu, A. Lukefahr, D. Palframan, G. Dasika, R. Das, and S. Mahlke, "Scalpel: Customizing DNN pruning to the underlying hardware parallelism," in Proceedings of International Symposium on Computer Architecture, pp. 548-560, 2017.
- P. Molchanov, S. Tyree, T. Karras, T. Aila, and J. Kautz, "Pruning convolutional neural networks for resource efficient inference," in Proceedings of International Conference on Learning Representations, pp. 1-17, 2017.
- T. Chen, Z. Du, N. Sun, J. Wang, C. Wu, Y. Chen, and O. Temam, "DianNao: A small-footprint high-throughput accelerator for ubiquitous machine-learning," in Proceedings of International Conference on Architectural Support for Programming Languages and Operating Systems, pp. 269-283, 2014.
- Y. Chen, T. Luo, S. Liu, S. Zhang, L. He, J. Wang, L. Li, T. Chen, Z. Xu, N. Sun, and O. Temam, "DaDianNao: A machine-learning supercomputer," in Proceedings of International Symposium on Micro-architecture, pp. 609-622, 2014.
- S. Liu, Z. Du, J. Tao, D. Han, T. Luo, Y. Zie, Y. Chen, and T. Chen, "Cambricon: An instruction set architecture for neural networks," in Proceedings of International Symposium on Computer Architecture, pp. 393-405, 2016.
- C. Zhang, P. Li, G. Sun, Y. Guan, B. Xiao, and J. Cong, "Optimizing FPGA-based accelerator design for deep convolutional neural networks," in Proceedings of ACM/SIGDA International Symposium on Field Programmable Gate Arrays, pp. 161-170, 2015.
- J. Albericio, P. Judd, T. Hetherington, T. Aamodt, N. E. Jerger, and A. Moshovos, "Cnvlutin: Ineffectual-neuron-free deep neural network computing," in Proceedings of International Symposium on Computer Architecture, pp. 1-13, 2016.
- L. Song, Y. Wang, Y. Han, X. Zhao, B. Liu, and X. Li, "C-Brain: A deep learning accelerator that tames the diversity of CNNs through adaptive data-level parallelization," in Proceedings of Design Automation Conference, pp. 123:1-123:6, 2016.
- J. Qiu, J. Wang, S. Yao, K. Guo, B. Li, E. Zhou, J. Yu, T. Tang, N. Xu, S. Song, Y. Wang, and H. Yang, "Going deeper with embedded FPGA platform for convolutional neural network," in Proceedings of ACM/SIGDA International Symposium on Field Programmable Gate Arrays, pp. 26-35, 2016.
- M. Motamedi, P. Gysel, V. Akella, and S. Ghiasi, "Design space exploration of FPGA-based deep convolutional neural networks," in Proceedings of Asia and South Pacific Design Automation Conference, pp. 575-580, 2016.
- J. Jo, S. Cha, D. Rho, and I.-C. Park, "DSIP: A scalable inference accelerator for convolutional neural networks," IEEE Journal of Solid-State Circuits, vol. 53, no. 2, pp. 605-618, Feb. 2018. https://doi.org/10.1109/JSSC.2017.2764045
- S. Han, X. Liu, H. Mao, J. Pu, A. Pedram, M. A. Horowitz, and W. J. Dally, "EIE: Efficient inference engine on compressed deep neural network," in Proceedings of International Symposium on Computer Architecture, pp. 243-254, 2016.
- S. Han, J. Kang, H. Mao, Y. Hu, X. Li, Y. Li, D. Xie, H. Luo, S. Yao, Y. Wang, H. Yang, and W. J. Dally, "ESE: Efficient speech recognition engine with sparse LSTM on FPGA," in Proceedings of ACM/SIGDA International Symposium on Field Programmable Gate Arrays, pp. 75-84, 2017.
- S. Zhang, Z. Du, L. Zhang, H. Lan, S. Liu, L. Li, Q. Guo, T. Chen, and Y. Chyen, "Cambricon-X: An accelerator for sparse neural networks," in Proceedings of International Symposium on Micro-architecture, pp. 20:1-20:12, 2016.
- D. Kim, J. Ahn, and S. Yoo, "A novel zero weight/ activation-aware hardware architecture of convolutional neural network," in Proceedings of Design, Automation & Test in Europe Conference & Exhibition, pp. 1462-1467, 2017.
- V. Lebedev and V. Lempitsky, "Fast ConvNets using group-wise brain damage," in Proceedings of Computer Vision and Pattern Recognition, pp. 2554-2564, 2016.
- S. Anwar, K. Hwang, and W. Sung, "Structured pruning of deep convolutional neural networks," ACM Journal on Emerging Technologies in Computing System, vol. 13, no. 3, pp. 32:1-32:18, Feb. 2017.