DOI QR코드

DOI QR Code

Development of an Efficient 3D Object Recognition Algorithm for Robotic Grasping in Cluttered Environments

혼재된 환경에서의 효율적 로봇 파지를 위한 3차원 물체 인식 알고리즘 개발

  • Received : 2022.05.24
  • Accepted : 2022.07.06
  • Published : 2022.08.31

Abstract

3D object detection pipelines often incorporate RGB-based object detection methods such as YOLO, which detects the object classes and bounding boxes from the RGB image. However, in complex environments where objects are heavily cluttered, bounding box approaches may show degraded performance due to the overlapping bounding boxes. Mask based methods such as Mask R-CNN can handle such situation better thanks to their detailed object masks, but they require much longer time for data preparation compared to bounding box-based approaches. In this paper, we present a 3D object recognition pipeline which uses either the YOLO or Mask R-CNN real-time object detection algorithm, K-nearest clustering algorithm, mask reduction algorithm and finally Principal Component Analysis (PCA) alg orithm to efficiently detect 3D poses of objects in a complex environment. Furthermore, we also present an improved YOLO based 3D object detection algorithm that uses a prioritized heightmap clustering algorithm to handle overlapping bounding boxes. The suggested algorithms have successfully been used at the Artificial-Intelligence Robot Challenge (ARC) 2021 competition with excellent results.

Keywords

Acknowledgement

This work was partially supported by IITP grant funded by the Korea government (MSIT) (No. 2021-0-01202) and KIAT grant funded by the Korea Government (MOTIE) (P0008473, HRD Program for Industrial Innovation) This project was funded by Police-Lab 2.0 Program (www.kipot.or.kr) funded by the Ministry of Science and ICT(MSIT, Korea) & Korean National Police Agency (KNPA, Korea) (No. 082021D48000000) and This work was supported by Institute of Information & communications Technology Planning & Evaluation (IITP) grant funded by the Korea government (MSIT) (No. 2021-0-01202)

References

  1. A . Bochkovskiy, C.-Y. Wang, and H.-Y. M. Liao, "YOLOv4: Optimal Speed and Accuracy of Object Detection," arXiv:2004.10934, 2020, DOI: 10.48550/arXiv.2004.10934.
  2. K. He, G. Gkioxari, P. Dollar, and R. Girshick, "Mask R-CNN," 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 2017, DOI: 10.1109/ICCV.2017.322.
  3. Y . Yoo, C.-Y. Lee, and B.-T. Zhang, "Multimodal Anomaly Detection based on Deep Auto-Encoder for Object Slip Perception of Mobile Manipulation Robots," 2021 IEEE International Conference on Robotics and Automation (ICRA), Xi'an, China, 2021, DOI: 10.1109/ICRA48506.2021.9561586.
  4. W . Kehl, F. Manhardt, F. Tombari, S. Ilic, and N. Navab, "SSD-6D: Making RGB-Based 3D Detection and 6D Pose Estimation Great Again," arXiv:1711.10006, 2017, DOI: 10.48550/arXiv.1711.10006.
  5. T .-Y. Lin, M. Maire, S. Belongie, J. Hays, P. Perona, D. Ramanan, P. Dollar, and C. L. Zitnick, "Microsoft COCO: Common Objects in Context," European Conference on Computer Vision, pp. 740-755, 2014, DOI: 10.1007/978-3-319-10602-1_48.
  6. D . Muller, N. Y. Wettengel, and D. Paulus, "homer@UniKoblenz: Winning Team of the RoboCup Virtual @Home Open Platform League 2021," RoboCup 2021: Robot World Cup XXIV, 2022, DOI: 10.1007/978-3-030-98682-7_23.
  7. M . F. B. van der Burgh, J. J. M. Lunenburg, R. P. W. Appeldoorn, L. L. A. M. van Beek, J. Geijsberts, L. G. L. Janssen, P. van Dooren, H. W. A. M. van Rooy, A. Aggarwal, S. Aleksandrov, K. Dang, A. T. Hofkamp, D. van Dinther, and M. J. G. van de Molengraft, "Tech United Eindhoven @Home2019 Champions Paper," RoboCup 2019: Robot World Cup XXIII, 2019, DOI: 10.1007/978-3-030-35699-6_43.
  8. S . G. McGill, S.-J. Yi, H. Yi, M. S. Ahn, S. Cho, K. Liu, D. Sun, B. Lee, H. Jeong, J. Huh, D. Hong, and D. D. Lee, "Team THOR's Entry in theDARPA Robotics Challenge Finals 2015," Journal of Field Robotics, vol. 34, no. 4, 2017, DOI: 10.1002/rob.21672.
  9. J .-B. Yi, T. Kang, D. Song, and S.-J. Yi, "Unified software platform for intelligent home service robots," Applied Sciences, vol. 10, no. 17, 2020, DOI: 10.3390/app10175874.
  10. S . Zakharov, I . Shugurov, and S. Ilic, "DPOD: 6D Pose Object Detector and Refiner," Computer Vision and Pattern Recognition, 2019, DOI: 10.48550/arXiv.1902.11020.
  11. H . Sun, Z. Meng, X. Du, and M. H. Ang, "A 3D Convolutional Neural Network Towards Real-Time Amodal 3D Object Detection," 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Madrid, Spain, 2018, DOI: 10.1109/IROS.2018.8593837.
  12. B . Tekin, S. N. Sudipta, and P. Fua, "Real-time seamless single shot 6d object pose prediction," Computer Vision and Pattern Recognition, 2018, DOI: 10.48550 /arXiv.1711.08848. https://doi.org/10.48550
  13. Y . Tanaka, Y. sIhida, Y. Abe, T. Ono, K. Kabashima, T. Sakata, M. Fukuyado, F. Muto, T. Yoshii, K. Kanamaru, D. Kamimura, K. Nakamura, Y. Nishimura, T. Morie, and H. Tamukoh, "Hibikino-Musashi@Home 2019 Team Description Paper," arXiv:2006.01233, 2019, DOI: 10.48550/arXiv.2006.01233.
  14. 2019 - TDP - Tech United Eindhoven - RoboCup@Home - Domestic Standard Platform, [Online], https://tdp.robocup.org/tdp/2019-tdp-tech-united-eindhoven-robocuphome-domestic-standard-platform/, Accessed: 03 16, 2022.
  15. R . Adams and L. Bischof, "Seeded region growing," IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 16, no. 6, pp. 641-647, June, 1994, DOI: 10.1109/34.295913.
  16. J . Fa n, D. K. Y. Yau, A. K. Elmagarmid, and W. G. Aref, "Automatic image segmentation by integrating color-edge extraction and seeded region growing," IEEE Transactions on Image Processing, vol. 10, no. 10, pp. 1454-1466, Oct., 2001, DOI: 10.1109/83.951532.
  17. H . Abdi and L. J. Williams, "Principal component analysis," WIREs Computational Statistics, vol. 2, no. 4, pp. 433-459, 2010, DOI: 10.1002/wics.101..
  18. T . Kang, .J-B. Yi, D. oSng, and S.-J. Yi, H"igh-speed uatonomous robotic assembly using in-handmanipulation and re-grasping," Applied Sciences, vol. 11, no. 1, 2021, DOI: 10.3390/app11010037.