Browse > Article
http://dx.doi.org/10.3837/tiis.2020.02.008

Fast, Accurate Vehicle Detection and Distance Estimation  

Ma, QuanMeng (School of Telecommunication Engineering, Xidian University)
Jiang, Guang (School of Telecommunication Engineering, Xidian University)
Lai, DianZhi (School of Telecommunication Engineering, Xidian University)
cui, Hua (School of information engineering, Chang'an University)
Song, Huansheng (School of information engineering, Chang'an University)
Publication Information
KSII Transactions on Internet and Information Systems (TIIS) / v.14, no.2, 2020 , pp. 610-630 More about this Journal
Abstract
A large number of people suffered from traffic accidents each year, so people pay more attention to traffic safety. However, the traditional methods use laser sensors to calculate the vehicle distance at a very high cost. In this paper, we propose a method based on deep learning to calculate the vehicle distance with a monocular camera. Our method is inexpensive and quite convenient to deploy on the mobile platforms. This paper makes two contributions. First, based on Light-Head RCNN, we propose a new vehicle detection framework called Light-Car Detection which can be used on the mobile platforms. Second, the planar homography of projective geometry is used to calculate the distance between the camera and the vehicles ahead. The results show that our detection system achieves 13FPS detection speed and 60.0% mAP on the Adreno 530 GPU of Samsung Galaxy S7, while only requires 7.1MB of storage space. Compared with the methods existed, the proposed method achieves a better performance.
Keywords
Light-Car Detection; Deep learning; vehicle distance; object detection; mobile platform;
Citations & Related Records
연도 인용수 순위
  • Reference
1 F. Tosi, F. Aleotti, M. Poggi, and S. Mattoccia. "Learning monocular depth estimation infusing traditional stereo knowledge," in Proc. of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 9799-9809, 2019.
2 H. Kong, J. Y. Audibert, and J. Pone, "Vanishing point detection for road detection," in Proc. of 2009 IEEE Conference on Computer Vision and Pattern Recognition, pp. 96-103, 2009.
3 C. H. Chen, T. Y. Chen, D. Y. Huang, and K. W. Feng, "Front vehicle detection and distance estimation using single-lens video camera," in Proc. of 2015 Third International Conference on Robot, Vision and Signal Processing (RVSP), pp. 14-17, 2015.
4 D. Y. Huang, C. H. Chen, T. Y. Chen, W. C. Hu, and K. W. Feng, "Vehicle detection and inter-vehicle distance estimation using single-lens video camera on urban/suburb roads," Journal Visual Communication and Image Representation, vol. 46, pp. 250-259, 2017.   DOI
5 A. Geiger, M. Lauer, C. Wojek, C. Stiller, and R. Urtasum, "3d traffic scene understanding from movable platforms," IEEE transactions on pattern analysis and machine intelligence, vol. 36, no. 5, pp. 1012-1025, 2014.   DOI
6 C. Chen, A. Seff, A. Kornhauser, and J. Xiao, "Deepdriving: Learning affordance for direct perception in autonomous driving," in Proc. of the IEEE International Conference on Computer Vision, pp. 2722-2730, 2015.
7 C. C. Tsai, Y. T. Lai, Y. F. Li, and J. G. Guo, "A vision radar system for car safety driving applications," in Proc. of 2017 International Symposium on VLSI Design, Automation and Test (VLSI-DAT), pp. 1-4, 2017.
8 M. Chen, D. Zhao, J. Sun, and H. Peng, "Improving Localization Accuracy in Connected Vehicle Networks Using Rao-Blackwellized Particle Filters: Theory, Simulations, and Experiments," IEEE Transactions on Intelligent Transportation Systems, vol.20, no.6, pp.2255-2266, 2019.   DOI
9 N. Sasaki, S. Tomaru, and S. Nakamura, "Development of inter-vehicle distance measurement system using camera-equipped portable device," in Proc. of 2017 17th International Conference on Control, Automation and Systems (ICCAS), pp. 994-997, 2017.
10 F. de Ponte Muller, "Survey on Ranging Sensors and Cooperative Techniques for Relative Positioning of Vehicles," Sensors, 17(2), 271, 2017.   DOI
11 K. Y. Park and S. Y. Hwang, "Robust Range Estimation with a Monocular Camera for Vision-Based Forward Collision Warning System," the Scientific World Journal, vol.2014, p.9, 2014.
12 A. Krizhevsky, I. Sutskever and G. E. Hinton, "Imagenet classification with deep convolutional neural networks," Advances in neural information processing systems, pp. 1097-1105, 2012.
13 K. Simonyan and A. Zisserman, "Very deep convolutional networks for large-scale image recognition," arXiv preprint arXiv:1409.1556, 2014.
14 C. Szegedy, W. Liu, Y. Jia, P. Sermanet, S. Reed, D. Anguelov, D. Erhan, V. Vanhoucke and A. Rabinovich, "Going deeper with convolutions," in Proc. of the IEEE conference on computer vision and pattern recognition, pp. 1-9, 2015.
15 M. Sandler, A. Howard, M. Zhu, A. Zhmoginov and L. C. Chen, "Inverted Residuals and Linear Bottlenecks: Mobile Networks for Classification, Detection and Segmentation," in Proc. of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 4510-4520, 2018.
16 K. He, X. Zhang, S. Ren and J. Sun, "Deep residual learning for image recognition," in Proc. of the IEEE conference on computer vision and pattern recognition, pp. 770-778, 2016.
17 S. Ioffe and C. Szegedy, "Batch normalization: Accelerating deep network training by reducing internal covariate shift," arXiv preprint arXiv:1502.03167, 2015.
18 S. Xie, R. Girshick, P. Dollar, Z. Tu and K. He, "Aggregated residual transformations for deep neural network," in Proc. of the IEEE conference on computer vision and pattern recognition, pp. 1492-1500, 2017.
19 J. Hu, L. Shen and G. Sun, "Squeeze-and-excitation networks," in Proc. of the IEEE conference on computer vision and pattern recognition, pp. 7132-7141, 2018.
20 F. N. Iandola, S. Han, M. W. Moskewicz, K. Ashraf, W. J. Dally and K. Keutzer, "SqueezeNet: AlexNet-level accuracy with 50x fewer parameters and< 0.5 MB model size," arXiv preprint arXiv:1602.07360, 2016.
21 W. Liu, D. Anguelov, D. Erhan, C. Szegedy, S. Reed, C. Y. Fu, and A. C. Berg, "Ssd: Single shot multibox detector," in Proc. of European conference on computer vision, pp. 21-37, 2016.
22 J. Redmon and A. Farhadi, "YOLO9000: better, faster, stronger," in Proc. of the IEEE conference on computer vision and pattern recognition, pp. 7263-7271, 2017.
23 S. Ren, K. He, R. Girshick and J. Sun, "Faster r-cnn: Towards real-time object detection with region proposal networks," Advances in neural information processing systems, pp. 91-99, 2015.
24 T. Y. Lin, P. Goyal, R. Girshick, K. He and P. Dollar, "Focal loss for dense object detection," in Proc. of the IEEE international conference on computer vision, pp. 2980-2988, 2017.
25 Redmon J, Farhadi A. "Yolov3: An incremental improvement," arXiv preprint arXiv:1804.02767, 2018.
26 J. Dai, Y. Li, K. He and J. Sun, "R-fcn: Object detection via region-based fully convolutional networks," Advances in neural information processing systems, pp. 379-387, 2016.
27 A. G. Howard, M. Zhu, B. Chen, D. Kalenichenko, W. Wang, T. Weyand, M. Andreetto and H. Adam, "Mobilenets: Efficient convolutional neural networks for mobile vision applications," arXiv preprint arXiv:1704.04861, 2017.
28 T. Y. Lin, M. Maire, S. Belongie, J. Hays, P. Perona, D. Ramanan, P. Dollar and C. L. Zitnick, "Microsoft coco: Common objects in context," in Proc. of European conference on computer vision, pp. 740-755, 2014.
29 A. Shrivastava, A. Gupta and R. Girshick, "Training region-based object detectors with online hard example mining," in Proc. of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 761-769, 2016.
30 Y. LeCun, B. Boser, J. S. Denker, D. Henderson, R. E. Howard, W. Hubbard and L. D. Jackel, "Backpropagation applied to handwritten zip code recognition," Neural computation, vol. 1, no. 4, pp. 541-551, 1989.   DOI
31 N. Bodla, B. Singh, R. Chellappa and L. S. Davis, "Soft-NMS--Improving Object Detection With One Line of Code," in Proc. of the IEEE International Conference on Computer Vision, pp. 5561-5569, 2017.
32 Yihui He, X. Zhang, J. Sun, "Channel pruning for accelerating very deep neural networks," in Proc. of The IEEE International Conference on Computer Vision (ICCV), pp. 1389-1397, 2017.
33 H. Zhang and N. Wang, "On The Stability of Video Detection and Tracking," arXiv preprint arXiv:1611.06467, 2016.
34 S. Gidaris and N. Komodakis, "Object detection via a multi-region and semantic segmentation-aware cnn model," in Proc. of the IEEE International Conference on Computer Vision, pp. 1134-1142, 2015.
35 Z. Li, C. Peng, G. Yu, X. Zhang, Y. Deng and J. Sun, "Light-Head R-CNN: In Defense of Two-Stage Object Detector," arXiv preprint arXiv:1711.07264, 2017.
36 Z. Kalal, K. Mikolajczyk and J. Matas, "Forward-backward error: Automatic detection of tracking failures," in Proc. of 2010 20th International Conference on Pattern Recognition, pp. 2756-2759, 2010.
37 D.A. Forsyth and J. Ponce, "Computer Vision: A Modern Approach," Pearson Education Inc. 2003.
38 H. Li, A. Kadav, I. Durdanovic, H. Samet and H. P. Graf, "Pruning filters for efficient convnets," arXiv preprint arXiv:1608.08710, 2016.
39 S. Han, H. Mao, W. J. Dally, "Deep compression: Compressing deep neural networks with pruning, trained quantization and huffman coding," arXiv preprint arXiv:1510.00149, 2015.
40 Y. Cuiping, "Research on Safe Distance between Vehicles for Freeway Anti-collision System," Process Automation Instrumentation, vol. 9, p. 005, 2008.
41 D. Eigen, C. Puhrsch, and R. Fergus. "Depth map prediction from a single image using a multi-scale deep network," Advances in neural information processing systems, pp. 2366-2374, 2014.
42 C. Godard. O. M. Aodha, and G. Brostow. "Digging into self-supervised monocular depth estimation," arXiv preprint arXiv:1806.01260. 2018.
43 H. Fu, M. Gong, C. Wang, K. Batmanghelich, and D. Tao. "Deep ordinal regression network for monocular depth estimation," in Proc. of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2002-2011, 2018.
44 Y. Kuznietsov, J. Stuckler, and B. Leibe. "Semi-supervised deep learning for monocular depth map prediction," in Proc. of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 6647-6655, 2017.