DOI QR코드

DOI QR Code

Detecting Vehicles That Are Illegally Driving on Road Shoulders Using Faster R-CNN

Faster R-CNN을 이용한 갓길 차로 위반 차량 검출

  • Go, MyungJin (NC SOFT AI Biz Division) ;
  • Park, Minju (Department of Big Data Application, Hannam University) ;
  • Yeo, Jiho (Department of Big Data Application, Hannam University)
  • 고명진 (엔씨소프트 AI Biz실) ;
  • 박민주 (한남대학교 빅데이터응용학과) ;
  • 여지호 (한남대학교 빅데이터응용학과)
  • Received : 2021.11.03
  • Accepted : 2022.02.14
  • Published : 2022.02.28

Abstract

According to the statistics about the fatal crashes that have occurred on the expressways for the last 5 years, those who died on the shoulders of the road has been as 3 times high as the others who died on the expressways. It suggests that the crashes on the shoulders of the road should be fatal, and that it would be important to prevent the traffic crashes by cracking down on the vehicles intruding the shoulders of the road. Therefore, this study proposed a method to detect a vehicle that violates the shoulder lane by using the Faster R-CNN. The vehicle was detected based on the Faster R-CNN, and an additional reading module was configured to determine whether there was a shoulder violation. For experiments and evaluations, GTAV, a simulation game that can reproduce situations similar to the real world, was used. 1,800 images of training data and 800 evaluation data were processed and generated, and the performance according to the change of the threshold value was measured in ZFNet and VGG16. As a result, the detection rate of ZFNet was 99.2% based on Threshold 0.8 and VGG16 93.9% based on Threshold 0.7, and the average detection speed for each model was 0.0468 seconds for ZFNet and 0.16 seconds for VGG16, so the detection rate of ZFNet was about 7% higher. The speed was also confirmed to be about 3.4 times faster. These results show that even in a relatively uncomplicated network, it is possible to detect a vehicle that violates the shoulder lane at a high speed without pre-processing the input image. It suggests that this algorithm can be used to detect violations of designated lanes if sufficient training datasets based on actual video data are obtained.

최근 5년간 고속도로에서 발생한 사망 사고의 통계를 살펴보면, 고속도로 전체 사망자 중 갓길에서 발생한 사망자의 사망률이 약 3배 높은 것으로 나타났다. 이는 갓길 사고 발생 시 사고의 심각도가 매우 높다는 것을 보여주며, 갓길 차로 위반 차량을 단속하여 사고를 미연에 방지하는 것이 중요하다는 것을 시시한다. 이에 본 연구는 Faster R-CNN 기법을 활용하여 갓길 차로 위반 차량을 검출할 수 있는 방법을 제안하였다. Faster R-CNN 기법을 기반으로 차량을 탐지하고, 추가적인 판독 모듈을 구성하여 갓길 위반 여부를 판단하였다. 실험 및 평가를 위해 현실세계와 유사하게 상황을 재현할 수 있는 시뮬레이션 게임인 GTAV를 활용하였다. 이미지 형태의 학습데이터 1,800장과 평가데이터 800장을 가공 및 생성하였으며, ZFNet과 VGG16에서 Threshold 값의 변화에 따른 성능을 측정하였다. 그 결과 Threshold 0.8 기준 ZFNet 99.2%, Threshold 0.7 기준 VGG16 93.9%의 검출율을 보였고, 모델 별 평균 검출 속도는 ZFNet 0.0468초, VGG16 0.16초를 기록하여 ZFNet의 검출율이 약 7% 정도 높았으며, 검출 속도 또한 약 3.4배 빠름을 확인하였다. 이는 비교적 복잡하지 않은 네트워크에서도 입력 영상의 전처리 없이 빠른 속도로 갓길 차로 위반 차량의 검출이 가능함을 보여주며, 실제 영상자료 기반의 학습데이터셋을 충분히 확보한다면 지정 차로 위반 검출에 본 알고리즘을 활용할 수 있다는 것을 시사한다.

Keywords

Acknowledgement

이 논문은 2021학년도 한남대학교 학술연구비 지원에 의하여 연구되었습니다.

References

  1. Alexander B., Script Hook V + Native Trainer, https://ko.gta5-mods.com/tools/script-hook-v, 2021.10.15.
  2. CamxxCore, Scripted Camera Tool 1.2.2, https://ko.gta5-mods.com/scripts/scripted-camera-tool-1-0, 2021.10.25.
  3. Chaudhari R., Dubey S., Kathale J. and Rao R.(2018), "Autonomous Driving Car Using Convolutional Neural Networks," In 2018 Second International Conference on Inventive Communication and Computational Technologies (ICICCT), IEEE, pp.936-940.
  4. Dalal N. and Triggs B.(2005), "Histograms of oriented gradients for human detection," IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'05), vol. 1, pp.886-893.
  5. David S., https://medium.com/self-driving-cars/term-1-in-depth-on-udacitys-self-driving-car-curriculum-ffcf46af0c08, 2021.10.25.
  6. Everingham M., Van Gool L., Williams C. K., Winn J. and Zisserman A.(2010), "The pascal visual object classes (voc) challenge," International Journal of Computer Vision, vol. 88, no. 2, pp.303-338. https://doi.org/10.1007/s11263-009-0275-4
  7. Girshick R.(2015), "Fast R-CNN," IEEE International Conference on Computer Vision (ICCV), pp.1440-1148.
  8. Girshick R., Donahue J., Darrell T. and Malik J.(2014), "Rich feature hierarchies for accurate object detection and semantic segmentation," Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp.580-587.
  9. He K., Zhang X., Ren S. and Sun J.(2015), "Spatial pyramid pooling in deep convolutional networks for visual recognition," IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 37, no. 9, pp.1904-1916. https://doi.org/10.1109/TPAMI.2015.2389824
  10. Jo J. M.(2017), "Highway drone patrol network topology and performance analysis for traffic violation enforcement," The Journal of the Korea Institute of Electronic Communication Sciences, vol. 12, no. 6, pp.1043-1048. https://doi.org/10.13067/JKIECS.2017.12.6.1043
  11. Johnson-Roberson M., Barto C., Mehta R., Sridhar S. N., Rosaen K. and Vasudevan R.(2016), Driving in the matrix: Can virtual worlds replace human-generated annotations for real world tasks?, arXiv preprint arXiv:1610.01983.
  12. Korea Expressway Corporation, http://data.ex.co.kr/portal/docu/docuList?datasetId=41, 2021.10.25.
  13. Korea Expressway Corporation, http://www.ex.co.kr/cmm/fms/FileDown.do?atchFileId=FILE_000000000013576&fileSn=0, 2021.10.25.
  14. Korea Road Traffic Authority(2012), Intervention and shoulder traffic violation enforcement equipment standard development research service report, pp.18-24.
  15. Krizhevsky A., Sutskever I. and Hinton G. E.(2012), "Imagenet classification with deep convolutional neural networks," Advances in Neural Information Processing Systems, vol. 25, pp.1097-1105.
  16. Lee Y. J., Kim S. J., Park G. M. and Park G. M.(2020), "Comparison of number plate recognition performance of Synthetic number plate generator using 2D and 3D rotation," Proceedings of the Korean Society of Broadcast Engineers Conference, The Korean Institute of Broadcast and Media Engineers, pp. 232-235.
  17. Lowe D. G.(2004), "Distinctive image features from scale-invariant keypoints," International Journal of Computer Vision, vol. 60, no. 2, pp.91-110. https://doi.org/10.1023/B:VISI.0000029664.99615.94
  18. Lungaro P., Tollmar K., Saeik F., Gisbert C. M. and Dubus G.(2018), "Demonstration of a low-cost hyper-realistic testbed for designing future onboard experiences," Adjunct Proceedings of the 10th International Conference on Automotive User Interfaces and Interactive Vehicular Applications, pp.235-238.
  19. Palazzi A., Borghi G., Abati D., Calderara S. and Cucchiara R.(2017), "Learning to map vehicles into bird's eye view," International Conference on Image Analysis and Processing, Springer, Cham, pp.233-243.
  20. Ren S., He K., Girshick R. and Sun J.(2015), "Faster r-cnn: Towards real-time object detection with region proposal networks," Advances in Neural Information Processing Systems, vol. 28, pp.91-99.
  21. Richter S. R., Vineet V., Roth S. and Koltun V.(2016), "Playing for data: Ground truth from computer games," European Conference on Computer Vision, Springer, Cham, pp.102-118.
  22. Ruano, A.(2017), DeepGTAV: A plugin for GTAV that transforms it into a vision-based self-driving car research environment. https://github.com/ai-tor/DeepGTAV, 2021.10.25.
  23. Rumelhart D. E., Hinton G. E. and Williams R. J.(1986), "Learning representations by back-propagating errors," Nature, vol. 323, no. 6088, pp.533-536. https://doi.org/10.1038/323533a0
  24. Simonyan K. and Zisserman A.(2014), Very deep convolutional networks for large-scale image recognition, arXiv preprint arXiv:1409.1556.
  25. Uijlings J. R., Van De Sande K. E., Gevers T. and Smeulders A. W.(2013), "Selective search for object recognition," International Journal of Computer Vision, vol. 104, no. 2, pp.154-171. https://doi.org/10.1007/s11263-013-0620-5
  26. Wang D., Devin C., Cai Q. Z., Yu F. and Darrell T.(2019), "Deep object-centric policies for autonomous driving," IEEE 2019 International Conference on Robotics and Automation (ICRA), pp.8853-8859.
  27. Yue X., Wu B., Seshia S. A., Keutzer K. and Sangiovanni-Vincentelli A. L.(2018), "A lidar point cloud generator: From a virtual world to autonomous driving," Proceedings of the 2018 ACM on International Conference on Multimedia Retrieval, pp.458-464.
  28. Zeiler M. D. and Fergus R.(2014) "Visualizing and understanding convolutional neural networks," European Conference on Computer Vision (ECCV), pp.818-833.