Browse > Article
http://dx.doi.org/10.7471/ikeee.2022.26.3.408

3D Point Cloud Reconstruction Technique from 2D Image Using Efficient Feature Map Extraction Network  

Kim, Jeong-Yoon (Dept. Electronic Engineering, Hanbat National University)
Lee, Seung-Ho (Dept. Electronic Engineering, Hanbat National University)
Publication Information
Journal of IKEEE / v.26, no.3, 2022 , pp. 408-415 More about this Journal
Abstract
In this paper, we propose a 3D point cloud reconstruction technique from 2D images using efficient feature map extraction network. The originality of the method proposed in this paper is as follows. First, we use a new feature map extraction network that is about 27% efficient than existing techniques in terms of memory. The proposed network does not reduce the size to the middle of the deep learning network, so important information required for 3D point cloud reconstruction is not lost. We solved the memory increase problem caused by the non-reduced image size by reducing the number of channels and by efficiently configuring the deep learning network to be shallow. Second, by preserving the high-resolution features of the 2D image, the accuracy can be further improved than that of the conventional technique. The feature map extracted from the non-reduced image contains more detailed information than the existing method, which can further improve the reconstruction accuracy of the 3D point cloud. Third, we use a divergence loss that does not require shooting information. The fact that not only the 2D image but also the shooting angle is required for learning, the dataset must contain detailed information and it is a disadvantage that makes it difficult to construct the dataset. In this paper, the accuracy of the reconstruction of the 3D point cloud can be increased by increasing the diversity of information through randomness without additional shooting information. In order to objectively evaluate the performance of the proposed method, using the ShapeNet dataset and using the same method as in the comparative papers, the CD value of the method proposed in this paper is 5.87, the EMD value is 5.81, and the FLOPs value is 2.9G. It was calculated. On the other hand, the lower the CD and EMD values, the better the accuracy of the reconstructed 3D point cloud approaches the original. In addition, the lower the number of FLOPs, the less memory is required for the deep learning network. Therefore, the CD, EMD, and FLOPs performance evaluation results of the proposed method showed about 27% improvement in memory and 6.3% in terms of accuracy compared to the methods in other papers, demonstrating objective performance.
Keywords
Point Cloud; Feature Map; Reconstruction; Reparameterization Trick; Latent Vector; Deep Learning;
Citations & Related Records
연도 인용수 순위
  • Reference
1 Fan, Haoqiang, Hao Su, and Leonidas J. Guibas. "A point set generation network for 3d object reconstruction from a single image," Proceedings of the IEEE conference on computer vision and pattern recognition, 2016. DOI: 10.48550/arXiv.1612.00603   DOI
2 Choy, C. B., Xu, D., Gwak, J., Chen, K., & Savarese, S.. "3d-r2n2: A unified approach for single and multi-view 3d object reconstruction," In European conference on computer vision, pp. 628-644. 2016. DOI: 10.48550/arXiv.1604.00449   DOI
3 KINGMA, Diederik P.; WELLING, Max. "Autoencoding variational bayes," International Conference on Learning Representations (ICLR), 2014. DOI: 10.48550/arXiv.1312.6114   DOI
4 Higgins, I., Matthey, L., Pal, A., Burgess, C., Glorot, X., Botvinick, M., Lerchner, A. "beta-vae: Learning basic visual concepts with a constrained variational framework." 2016.
5 Mandikal, P., Navaneet, K. L., Agarwal, M., Babu, R. V. "3D-LMNet: Latent embedding matching for accurate and diverse 3D point cloud reconstruction from a single image," Proceedings of the British Machine Vision Conference(BMVC), 2018. DOI:10.48550/arXiv.1807.07796   DOI
6 Li, B., Zhang, Y., Zhao, B., & Shao, H. "3D-ReConstnet: a single-view 3d-object point cloud reconstruction network," IEEE Access 8 pp.83782-83790. 2020. DOI: 10.1109/ACCESS.2020.2992554   DOI
7 Chang, A. X., Funkhouser, T., Guibas, L., Hanrahan, P., Huang, Q., Li, Z., ... & Yu, F.. "Shapenet: An information-rich 3d model repository," 2015. DOI: 10.48550/arXiv.1512.03012   DOI
8 He, K., Zhang, X., Ren, S., & Sun, J. (2016). "Deep residual learning for image recognition," In Proceedings of the IEEE conference on computer vision and pattern recognition, pp.770-778. DOI: 10.48550/arXiv.1512.03385   DOI