• Title/Summary/Keyword: 3D Point Cloud

Search Result 410, Processing Time 0.031 seconds

Automatic Building Modeling Method Using Planar Analysis of Point Clouds from Unmanned Aerial Vehicles (무인항공기에서 생성된 포인트 클라우드의 평면성 분석을 통한 자동 건물 모델 생성 기법)

  • Kim, Han-gyeol;Hwang, YunHyuk;Rhee, Sooahm
    • Korean Journal of Remote Sensing
    • /
    • v.35 no.6_1
    • /
    • pp.973-985
    • /
    • 2019
  • In this paper, we propose a method to separate the ground and building areas and generate building models automatically through planarity analysis using UAV (Unmanned Aerial Vehicle) based point cloud. In this study, proposed method includes five steps. In the first step, the planes of the point cloud were extracted by analyzing the planarity of the input point cloud. In the second step, the extracted planes were analyzed to find a plane corresponding to the ground surface. Then, the points corresponding to the plane were removed from the point cloud. In the third step, we generate ortho-projected image from the point cloud ground surface removed. In the fourth step, the outline of each object was extracted from the ortho-projected image. Then, the non-building area was removed using the area, area / length ratio. Finally, the building's outer points were constructed using the building's ground height and the building's height. Then, 3D building models were created. In order to verify the proposed method, we used point clouds made using the UAV images. Through experiments, we confirmed that the 3D models of the building were generated automatically.

3D Point Cloud Reconstruction Technique from 2D Image Using Efficient Feature Map Extraction Network (효율적인 feature map 추출 네트워크를 이용한 2D 이미지에서의 3D 포인트 클라우드 재구축 기법)

  • Kim, Jeong-Yoon;Lee, Seung-Ho
    • Journal of IKEEE
    • /
    • v.26 no.3
    • /
    • pp.408-415
    • /
    • 2022
  • In this paper, we propose a 3D point cloud reconstruction technique from 2D images using efficient feature map extraction network. The originality of the method proposed in this paper is as follows. First, we use a new feature map extraction network that is about 27% efficient than existing techniques in terms of memory. The proposed network does not reduce the size to the middle of the deep learning network, so important information required for 3D point cloud reconstruction is not lost. We solved the memory increase problem caused by the non-reduced image size by reducing the number of channels and by efficiently configuring the deep learning network to be shallow. Second, by preserving the high-resolution features of the 2D image, the accuracy can be further improved than that of the conventional technique. The feature map extracted from the non-reduced image contains more detailed information than the existing method, which can further improve the reconstruction accuracy of the 3D point cloud. Third, we use a divergence loss that does not require shooting information. The fact that not only the 2D image but also the shooting angle is required for learning, the dataset must contain detailed information and it is a disadvantage that makes it difficult to construct the dataset. In this paper, the accuracy of the reconstruction of the 3D point cloud can be increased by increasing the diversity of information through randomness without additional shooting information. In order to objectively evaluate the performance of the proposed method, using the ShapeNet dataset and using the same method as in the comparative papers, the CD value of the method proposed in this paper is 5.87, the EMD value is 5.81, and the FLOPs value is 2.9G. It was calculated. On the other hand, the lower the CD and EMD values, the better the accuracy of the reconstructed 3D point cloud approaches the original. In addition, the lower the number of FLOPs, the less memory is required for the deep learning network. Therefore, the CD, EMD, and FLOPs performance evaluation results of the proposed method showed about 27% improvement in memory and 6.3% in terms of accuracy compared to the methods in other papers, demonstrating objective performance.

3D Shape Descriptor for Segmenting Point Cloud Data

  • Park, So Young;Yoo, Eun Jin;Lee, Dong-Cheon;Lee, Yong Wook
    • Journal of the Korean Society of Surveying, Geodesy, Photogrammetry and Cartography
    • /
    • v.30 no.6_2
    • /
    • pp.643-651
    • /
    • 2012
  • Object recognition belongs to high-level processing that is one of the difficult and challenging tasks in computer vision. Digital photogrammetry based on the computer vision paradigm has begun to emerge in the middle of 1980s. However, the ultimate goal of digital photogrammetry - intelligent and autonomous processing of surface reconstruction - is not achieved yet. Object recognition requires a robust shape description about objects. However, most of the shape descriptors aim to apply 2D space for image data. Therefore, such descriptors have to be extended to deal with 3D data such as LiDAR(Light Detection and Ranging) data obtained from ALS(Airborne Laser Scanner) system. This paper introduces extension of chain code to 3D object space with hierarchical approach for segmenting point cloud data. The experiment demonstrates effectiveness and robustness of the proposed method for shape description and point cloud data segmentation. Geometric characteristics of various roof types are well described that will be eventually base for the object modeling. Segmentation accuracy of the simulated data was evaluated by measuring coordinates of the corners on the segmented patch boundaries. The overall RMSE(Root Mean Square Error) is equivalent to the average distance between points, i.e., GSD(Ground Sampling Distance).

Generating a Rectangular Net from Unorganized Point Cloud Data Using an Implicit Surface Scheme (음 함수 곡면기법을 이용한 임의의 점 군 데이터로부터의 사각망 생성)

  • Yoo, D.J.
    • Korean Journal of Computational Design and Engineering
    • /
    • v.12 no.4
    • /
    • pp.274-282
    • /
    • 2007
  • In this paper, a method of constructing a rectangular net from unorganized point cloud data is presented. In the method an implicit surface that fits the given point data is generated by using principal component analysis(PCA) and adaptive domain decomposition method(ADDM). Then a complete and quality rectangular net can be obtained by extracting voxel data from the implicit surface and projecting exterior faces of extracted voxels onto the implicit surface. The main advantage of the proposed method is that a quality rectangular net can be extracted from randomly scattered 3D points only without any further information. Furthermore the results of this works can be used to obtain many useful information including a slicing data, a solid STL model and a NURBS surface model in many areas involved in treatment of large amount of point data by proper processing of implicit surface and rectangular net generated previously.

Matching for the Elbow Cylinder Shape in the Point Cloud Using the PCA (주성분 분석을 통한 포인트 클라우드 굽은 실린더 형태 매칭)

  • Jin, YoungHoon
    • Journal of KIISE
    • /
    • v.44 no.4
    • /
    • pp.392-398
    • /
    • 2017
  • The point-cloud representation of an object is performed by scanning a space through a laser scanner that is extracting a set of points, and the points are then integrated into the same coordinate system through a registration. The set of the completed registration-integrated point clouds is classified into meaningful regions, shapes, and noises through a mathematical analysis. In this paper, the aim is the matching of a curved area like a cylinder shape in 3D point-cloud data. The matching procedure is the attainment of the center and radius data through the extraction of the cylinder-shape candidates from the sphere that is fitted through the RANdom Sample Consensus (RANSAC) in the point cloud, and completion requires the matching of the curved region with the Catmull-Rom spline from the extracted center-point data using the Principal Component Analysis (PCA). Not only is the proposed method expected to derive a fast estimation result via linear and curved cylinder estimations after a center-axis estimation without constraint and segmentation, but it should also increase the work efficiency of reverse engineering.

Automated Derivation of Cross-sectional Numerical Information of Retaining Walls Using Point Cloud Data (점군 데이터를 활용한 옹벽의 단면 수치 정보 자동화 도출)

  • Han, Jehee;Jang, Minseo;Han, Hyungseo;Jo, Hyoungjun;Shin, Do Hyoung
    • Journal of KIBIM
    • /
    • v.14 no.2
    • /
    • pp.1-12
    • /
    • 2024
  • The paper proposes a methodology that combines the Random Sample Consensus (RANSAC) algorithm and the Point Cloud Encoder-Decoder Network (PCEDNet) algorithm to automatically extract the length of infrastructure elements from point cloud data acquired through 3D LiDAR scans of retaining walls. This methodology is expected to significantly improve time and cost efficiency compared to traditional manual measurement techniques, which are crucial for the data-driven analysis required in the precision-demanding construction sector. Additionally, the extracted positional and dimensional data can contribute to enhanced accuracy and reliability in Scan-to-BIM processes. The results of this study are anticipated to provide important insights that could accelerate the digital transformation of the construction industry. This paper provides empirical data on how the integration of digital technologies can enhance efficiency and accuracy in the construction industry, and offers directions for future research and application.

Progress Measurement of Structural Frame Construction using Point Cloud Data (포인트 클라우드 데이터를 활용한 골조공사 진도측정 연구)

  • Kim, Ju-Yong;Kim, Sanghee;Kim, Gwang-Hee
    • Korean Journal of Construction Engineering and Management
    • /
    • v.25 no.3
    • /
    • pp.37-46
    • /
    • 2024
  • Recently, 3D laser scanning technology, which can collect accurate and quick information on phenomena, has been attracting attention among smart construction technologies. 3D laser scanning technology can obtain information most similar to reality at construction sites. In this study, we would like to apply a new member identification method to an actual building and present the possibility of applying point cloud data, which can be collected using 3D laser scanning technology, to measuring progress at construction sites. In order to carry out the research, we collected location information for component identification from BIM, set a recognition margin for the collected location information, and proceeded to identify the components that make up the building from point cloud data. Research results We confirmed that the columns, beams, walls, and slabs that make up a building can be identified from point cloud data. The identification results can be used to confirm all the parts that have been completed in the actual building, and can be used in conjunction with the unit price of each part in the project BOQ for prefabricated calculations. In addition, the point cloud data obtained through research can be used as accurate data for quality control monitoring of construction sites and building maintenance management. The research results can contribute to improving the timeliness and accuracy of construction information used in future project applications.

Real-time 3D Volumetric Model Generation using Multiview RGB-D Camera (다시점 RGB-D 카메라를 이용한 실시간 3차원 체적 모델의 생성)

  • Kim, Kyung-Jin;Park, Byung-Seo;Kim, Dong-Wook;Kwon, Soon-Chul;Seo, Young-Ho
    • Journal of Broadcast Engineering
    • /
    • v.25 no.3
    • /
    • pp.439-448
    • /
    • 2020
  • In this paper, we propose a modified optimization algorithm for point cloud matching of multi-view RGB-D cameras. In general, in the computer vision field, it is very important to accurately estimate the position of the camera. The 3D model generation methods proposed in the previous research require a large number of cameras or expensive 3D cameras. Also, the methods of obtaining the external parameters of the camera through the 2D image have a large error. In this paper, we propose a matching technique for generating a 3D point cloud and mesh model that can provide omnidirectional free viewpoint using 8 low-cost RGB-D cameras. We propose a method that uses a depth map-based function optimization method with RGB images and obtains coordinate transformation parameters that can generate a high-quality 3D model without obtaining initial parameters.

Conversion Method of 3D Point Cloud to Depth Image and Its Hardware Implementation (3차원 점군데이터의 깊이 영상 변환 방법 및 하드웨어 구현)

  • Jang, Kyounghoon;Jo, Gippeum;Kim, Geun-Jun;Kang, Bongsoon
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.18 no.10
    • /
    • pp.2443-2450
    • /
    • 2014
  • In the motion recognition system using depth image, the depth image is converted to the real world formed 3D point cloud data for efficient algorithm apply. And then, output depth image is converted by the projective world after algorithm apply. However, when coordinate conversion, rounding error and data loss by applied algorithm are occurred. In this paper, when convert 3D point cloud data to depth image, we proposed efficient conversion method and its hardware implementation without rounding error and data loss according image size change. The proposed system make progress using the OpenCV and the window program, and we test a system using the Kinect in real time. In addition, designed using Verilog-HDL and verified through the Zynq-7000 FPGA Board of Xilinx.

Segmentation and Classification of Lidar data

  • Tseng, Yi-Hsing;Wang, Miao
    • Proceedings of the KSRS Conference
    • /
    • 2003.11a
    • /
    • pp.153-155
    • /
    • 2003
  • Laser scanning has become a viable technique for the collection of a large amount of accurate 3D point data densely distributed on the scanned object surface. The inherent 3D nature of the sub-randomly distributed point cloud provides abundant spatial information. To explore valuable spatial information from laser scanned data becomes an active research topic, for instance extracting digital elevation model, building models, and vegetation volumes. The sub-randomly distributed point cloud should be segmented and classified before the extraction of spatial information. This paper investigates some exist segmentation methods, and then proposes an octree-based split-and-merge segmentation method to divide lidar data into clusters belonging to 3D planes. Therefore, the classification of lidar data can be performed based on the derived attributes of extracted 3D planes. The test results of both ground and airborne lidar data show the potential of applying this method to extract spatial features from lidar data.

  • PDF