• Title/Summary/Keyword: Automatic Extraction of Control Points

Search Result 18, Processing Time 0.024 seconds

Building Roof Reconstruction in Remote Sensing Image using Line Segment Extraction and Grouping (선소의 추출과 그룹화를 이용한 원격탐사영상에서 건물 지붕의 복원)

  • 예철수;전승헌;이호영;이쾌희
    • Korean Journal of Remote Sensing
    • /
    • v.19 no.2
    • /
    • pp.159-169
    • /
    • 2003
  • This paper presents a method for automatic 3-d building reconstruction using high resolution aerial imagery. First, by using edge preserving filtering, noise is eliminated and then images are segmented by watershed algorithm, which preserves location of edge pixels. To extract line segments between control points from boundary of each region, we calculate curvature of each pixel on the boundary and then find the control points. Line segment linking is performed according to direction and length of line segments and the location of line segments is adjusted using gradient magnitudes of all pixels of the line segment. Coplanar grouping and pplygonal patch formation are performed per region by selecting 3-d line segments that are matched using epipolar geometry and flight information. The algorithm has been applied to high resolution aerial images and the results show accurate 3D building reconstruction.

Application Possibility of Control Points Extracted from Ortho Images and DTED Level 2 for High Resolution Satellite Sensor Modeling (정사영상과 DTED Level 2 자료에서 자동 추출한 지상기준점의 IKONOS 위성영상 모델링 적용 가능성 연구)

  • Lee, Tae-Yoon;Kim, Tae-Jung;Park, Wan-Yong
    • Journal of Korean Society for Geospatial Information Science
    • /
    • v.15 no.4
    • /
    • pp.103-109
    • /
    • 2007
  • Ortho images and Digital Elevation Model (DEM) have been applied in various fields. It is necessary to acquire Ground Control Points (GCPs) for processing high resolution satellite images. However surveying GCPs require many time and expense. This study was performed to investigate whether GCPs automatically extracted from ortho images and DTED Level 2 can be applied to sensor modeling for high resolution satellite images. We analyzed the performance of the sensor model established by GCPs extracted automatically. We acquired GCPs by matching satellite image against ortho images. We included the height acquired from DTED Level 2 data in these GCPs. The spatial resolution of the DTED Level 2 data is about 30m. Absolution accuracy of this data is below 18m above MSL. The spatial resolution of ortho image is 1m. We established sensor model from IKONOS images using GCPs extracted automatically and generated DEMs from the images. The accuracy of sensor modeling is about $4{\sim}5$ pixel. We also established sensor models using GCPs acquired based on GPS surveying and generated DEMs. Two DEMs were similar. The RMSE of height from the DEM by automatic GCPs and DTED Level 2 is about 9 m. So we think that GCPs by DTED Level 2 and ortho image can use for IKONOS sensor modeling.

  • PDF

RPC Correction of KOMPSAT-3A Satellite Image through Automatic Matching Point Extraction Using Unmanned AerialVehicle Imagery (무인항공기 영상 활용 자동 정합점 추출을 통한 KOMPSAT-3A 위성영상의 RPC 보정)

  • Park, Jueon;Kim, Taeheon;Lee, Changhui;Han, Youkyung
    • Korean Journal of Remote Sensing
    • /
    • v.37 no.5_1
    • /
    • pp.1135-1147
    • /
    • 2021
  • In order to geometrically correct high-resolution satellite imagery, the sensor modeling process that restores the geometric relationship between the satellite sensor and the ground surface at the image acquisition time is required. In general, high-resolution satellites provide RPC (Rational Polynomial Coefficient) information, but the vendor-provided RPC includes geometric distortion caused by the position and orientation of the satellite sensor. GCP (Ground Control Point) is generally used to correct the RPC errors. The representative method of acquiring GCP is field survey to obtain accurate ground coordinates. However, it is difficult to find the GCP in the satellite image due to the quality of the image, land cover change, relief displacement, etc. By using image maps acquired from various sensors as reference data, it is possible to automate the collection of GCP through the image matching algorithm. In this study, the RPC of KOMPSAT-3A satellite image was corrected through the extracted matching point using the UAV (Unmanned Aerial Vehichle) imagery. We propose a pre-porocessing method for the extraction of matching points between the UAV imagery and KOMPSAT-3A satellite image. To this end, the characteristics of matching points extracted by independently applying the SURF (Speeded-Up Robust Features) and the phase correlation, which are representative feature-based matching method and area-based matching method, respectively, were compared. The RPC adjustment parameters were calculated using the matching points extracted through each algorithm. In order to verify the performance and usability of the proposed method, it was compared with the GCP-based RPC correction result. The GCP-based method showed an improvement of correction accuracy by 2.14 pixels for the sample and 5.43 pixelsfor the line compared to the vendor-provided RPC. In the proposed method using SURF and phase correlation methods, the accuracy of sample was improved by 0.83 pixels and 1.49 pixels, and that of line wasimproved by 4.81 pixels and 5.19 pixels, respectively, compared to the vendor-provided RPC. Through the experimental results, the proposed method using the UAV imagery presented the possibility as an alternative to the GCP-based method for the RPC correction.

Omni-directional Vision SLAM using a Motion Estimation Method based on Fisheye Image (어안 이미지 기반의 움직임 추정 기법을 이용한 전방향 영상 SLAM)

  • Choi, Yun Won;Choi, Jeong Won;Dai, Yanyan;Lee, Suk Gyu
    • Journal of Institute of Control, Robotics and Systems
    • /
    • v.20 no.8
    • /
    • pp.868-874
    • /
    • 2014
  • This paper proposes a novel mapping algorithm in Omni-directional Vision SLAM based on an obstacle's feature extraction using Lucas-Kanade Optical Flow motion detection and images obtained through fish-eye lenses mounted on robots. Omni-directional image sensors have distortion problems because they use a fish-eye lens or mirror, but it is possible in real time image processing for mobile robots because it measured all information around the robot at one time. In previous Omni-Directional Vision SLAM research, feature points in corrected fisheye images were used but the proposed algorithm corrected only the feature point of the obstacle. We obtained faster processing than previous systems through this process. The core of the proposed algorithm may be summarized as follows: First, we capture instantaneous $360^{\circ}$ panoramic images around a robot through fish-eye lenses which are mounted in the bottom direction. Second, we remove the feature points of the floor surface using a histogram filter, and label the candidates of the obstacle extracted. Third, we estimate the location of obstacles based on motion vectors using LKOF. Finally, it estimates the robot position using an Extended Kalman Filter based on the obstacle position obtained by LKOF and creates a map. We will confirm the reliability of the mapping algorithm using motion estimation based on fisheye images through the comparison between maps obtained using the proposed algorithm and real maps.

Automation of Bio-Industrial Process Via Tele-Task Command(I) -identification and 3D coordinate extraction of object- (원격작업 지시를 이용한 생물산업공정의 생력화 (I) -대상체 인식 및 3차원 좌표 추출-)

  • Kim, S. C.;Choi, D. Y.;Hwang, H.
    • Journal of Biosystems Engineering
    • /
    • v.26 no.1
    • /
    • pp.21-28
    • /
    • 2001
  • Major deficiencies of current automation scheme including various robots for bioproduction include the lack of task adaptability and real time processing, low job performance for diverse tasks, and the lack of robustness of take results, high system cost, failure of the credit from the operator, and so on. This paper proposed a scheme that could solve the current limitation of task abilities of conventional computer controlled automatic system. The proposed scheme is the man-machine hybrid automation via tele-operation which can handle various bioproduction processes. And it was classified into two categories. One category was the efficient task sharing between operator and CCM(computer controlled machine). The other was the efficient interface between operator and CCM. To realize the proposed concept, task of the object identification and extraction of 3D coordinate of an object was selected. 3D coordinate information was obtained from camera calibration using camera as a measurement device. Two stereo images were obtained by moving a camera certain distance in horizontal direction normal to focal axis and by acquiring two images at different locations. Transformation matrix for camera calibration was obtained via least square error approach using specified 6 known pairs of data points in 2D image and 3D world space. 3D world coordinate was obtained from two sets of image pixel coordinates of both camera images with calibrated transformation matrix. As an interface system between operator and CCM, a touch pad screen mounted on the monitor and remotely captured imaging system were used. Object indication was done by the operator’s finger touch to the captured image using the touch pad screen. A certain size of local image processing area was specified after the touch was made. And image processing was performed with the specified local area to extract desired features of the object. An MS Windows based interface software was developed using Visual C++6.0. The software was developed with four modules such as remote image acquisiton module, task command module, local image processing module and 3D coordinate extraction module. Proposed scheme shoed the feasibility of real time processing, robust and precise object identification, and adaptability of various job and environments though selected sample tasks.

  • PDF

Automatic Detection of Malfunctioning Photovoltaic Modules Using Unmanned Aerial Vehicle Thermal Infrared Images

  • Kim, Dusik;Youn, Junhee;Kim, Changyoon
    • Journal of the Korean Society of Surveying, Geodesy, Photogrammetry and Cartography
    • /
    • v.34 no.6
    • /
    • pp.619-627
    • /
    • 2016
  • Cells of a PV (photovoltaic) module can suffer defects due to various causes resulting in a loss of power output. As a malfunctioning cell has a higher temperature than adjacent normal cells, it can be easily detected with a thermal infrared sensor. A conventional method of PV cell inspection is to use a hand-held infrared sensor for visual inspection. The main disadvantages of this method, when applied to a large-scale PV power plant, are that it is time-consuming and costly. This paper presents an algorithm for automatically detecting defective PV panels using images captured with a thermal imaging camera from an UAV (unmanned aerial vehicle). The proposed algorithm uses statistical analysis of thermal intensity (surface temperature) characteristics of each PV module to verify the mean intensity and standard deviation of each panel as parameters for fault diagnosis. One of the characteristics of thermal infrared imaging is that the larger the distance between sensor and target, the lower the measured temperature of the object. Consequently, a global detection rule using the mean intensity of all panels in the fault detection algorithm is not applicable. Therefore, a local detection rule was applied to automatically detect defective panels using the mean intensity and standard deviation range of each panel by array. The performance of the proposed algorithm was tested on three sample images; this verified a detection accuracy of defective panels of 97% or higher. In addition, as the proposed algorithm can adjust the range of threshold values for judging malfunction at the array level, the local detection rule is considered better suited for highly sensitive fault detection compared to a global detection rule. In this study, we used a panel area extraction method that we previously developed; fault detection accuracy would be improved if panel area extraction from images was more precise. Furthermore, the proposed algorithm contributes to the development of a maintenance and repair system for large-scale PV power plants, in combination with a geo-referencing algorithm for accurate determination of panel locations using sensor-based orientation parameters and photogrammetry from ground control points.

Development of an Image Processing System for the Large Size High Resolution Satellite Images (대용량 고해상 위성영상처리 시스템 개발)

  • 김경옥;양영규;안충현
    • Korean Journal of Remote Sensing
    • /
    • v.14 no.4
    • /
    • pp.376-391
    • /
    • 1998
  • Images from satellites will have 1 to 3 meter ground resolution and will be very useful for analyzing current status of earth surface. An image processing system named GeoWatch with more intelligent image processing algorithms has been designed and implemented to support the detailed analysis of the land surface using high-resolution satellite imagery. The GeoWatch is a valuable tool for satellite image processing such as digitizing, geometric correction using ground control points, interactive enhancement, various transforms, arithmetic operations, calculating vegetation indices. It can be used for investigating various facts such as the change detection, land cover classification, capacity estimation of the industrial complex, urban information extraction, etc. using more intelligent analysis method with a variety of visual techniques. The strong points of this system are flexible algorithm-save-method for efficient handling of large size images (e.g. full scenes), automatic menu generation and powerful visual programming environment. Most of the existing image processing systems use general graphic user interfaces. In this paper we adopted visual program language for remotely sensed image processing for its powerful programmability and ease of use. This system is an integrated raster/vector analysis system and equipped with many useful functions such as vector overlay, flight simulation, 3D display, and object modeling techniques, etc. In addition to the modules for image and digital signal processing, the system provides many other utilities such as a toolbox and an interactive image editor. This paper also presents several cases of image analysis methods with AI (Artificial Intelligent) technique and design concept for visual programming environment.

A Study on the Integration of Airborne LiDAR and UAV Data for High-resolution Topographic Information Construction of Tidal Flat (갯벌지역 고해상도 지형정보 구축을 위한 항공 라이다와 UAV 데이터 통합 활용에 관한 연구)

  • Kim, Hye Jin;Lee, Jae Bin;Kim, Yong Il
    • Journal of the Korean Society of Surveying, Geodesy, Photogrammetry and Cartography
    • /
    • v.38 no.4
    • /
    • pp.345-352
    • /
    • 2020
  • To preserve and restore tidal flats and prevent safety accidents, it is necessary to construct tidal flat topographic information including the exact location and shape of tidal creeks. In the tidal flats where the field surveying is difficult to apply, airborne LiDAR surveying can provide accurate terrain data for a wide area. On the other hand, we can economically obtain relatively high-resolution data from UAV (Unmanned Aerial Vehicle) surveying. In this study, we proposed the methodology to generate high-resolution topographic information of tidal flats effectively by integrating airborne LiDAR and UAV point clouds. For the purpose, automatic ICP (Iterative Closest Points) registration between two different datasets was conducted and tidal creeks were extracted by applying CSF (Cloth Simulation Filtering) algorithm. Then, we integrated high-density UAV data for tidal creeks and airborne LiDAR data for flat grounds. DEM (Digital Elevation Model) and tidal flat area and depth were generated from the integrated data to construct high-resolution topographic information for large-scale tidal flat map creation. As a result, UAV data was registered without GCP (Ground Control Point), and integrated data including detailed topographic information of tidal creeks with a relatively small data size was generated.