• Title/Summary/Keyword: Image segmentation

Search Result 2,165, Processing Time 0.029 seconds

Object-based Building Change Detection Using Azimuth and Elevation Angles of Sun and Platform in the Multi-sensor Images (태양과 플랫폼의 방위각 및 고도각을 이용한 이종 센서 영상에서의 객체기반 건물 변화탐지)

  • Jung, Sejung;Park, Jueon;Lee, Won Hee;Han, Youkyung
    • Korean Journal of Remote Sensing
    • /
    • v.36 no.5_2
    • /
    • pp.989-1006
    • /
    • 2020
  • Building change monitoring based on building detection is one of the most important fields in terms of monitoring artificial structures using high-resolution multi-temporal images such as CAS500-1 and 2, which are scheduled to be launched. However, not only the various shapes and sizes of buildings located on the surface of the Earth, but also the shadows or trees around them make it difficult to detect the buildings accurately. Also, a large number of misdetection are caused by relief displacement according to the azimuth and elevation angles of the platform. In this study, object-based building detection was performed using the azimuth angle of the Sun and the corresponding main direction of shadows to improve the results of building change detection. After that, the platform's azimuth and elevation angles were used to detect changed buildings. The object-based segmentation was performed on a high-resolution imagery, and then shadow objects were classified through the shadow intensity, and feature information such as rectangular fit, Gray-Level Co-occurrence Matrix (GLCM) homogeneity and area of each object were calculated for building candidate detection. Then, the final buildings were detected using the direction and distance relationship between the center of building candidate object and its shadow according to the azimuth angle of the Sun. A total of three methods were proposed for the building change detection between building objects detected in each image: simple overlay between objects, comparison of the object sizes according to the elevation angle of the platform, and consideration of direction between objects according to the azimuth angle of the platform. In this study, residential area was selected as study area using high-resolution imagery acquired from KOMPSAT-3 and Unmanned Aerial Vehicle (UAV). Experimental results have shown that F1-scores of building detection results detected using feature information were 0.488 and 0.696 respectively in KOMPSAT-3 image and UAV image, whereas F1-scores of building detection results considering shadows were 0.876 and 0.867, respectively, indicating that the accuracy of building detection method considering shadows is higher. Also among the three proposed building change detection methods, the F1-score of the consideration of direction between objects according to the azimuth angles was the highest at 0.891.

Automatic Fracture Detection in CT Scan Images of Rocks Using Modified Faster R-CNN Deep-Learning Algorithm with Rotated Bounding Box (회전 경계박스 기능의 변형 FASTER R-CNN 딥러닝 알고리즘을 이용한 암석 CT 영상 내 자동 균열 탐지)

  • Pham, Chuyen;Zhuang, Li;Yeom, Sun;Shin, Hyu-Soung
    • Tunnel and Underground Space
    • /
    • v.31 no.5
    • /
    • pp.374-384
    • /
    • 2021
  • In this study, we propose a new approach for automatic fracture detection in CT scan images of rock specimens. This approach is built on top of two-stage object detection deep learning algorithm called Faster R-CNN with a major modification of using rotated bounding box. The use of rotated bounding box plays a key role in the future work to overcome several inherent difficulties of fracture segmentation relating to the heterogeneity of uninterested background (i.e., minerals) and the variation in size and shape of fracture. Comparing to the commonly used bounding box (i.e., axis-align bounding box), rotated bounding box shows a greater adaptability to fit with the elongated shape of fracture, such that minimizing the ratio of background within the bounding box. Besides, an additional benefit of rotated bounding box is that it can provide relative information on the orientation and length of fracture without the further segmentation and measurement step. To validate the applicability of the proposed approach, we train and test our approach with a number of CT image sets of fractured granite specimens with highly heterogeneous background and other rocks such as sandstone and shale. The result demonstrates that our approach can lead to the encouraging results on fracture detection with the mean average precision (mAP) up to 0.89 and also outperform the conventional approach in terms of background-to-object ratio within the bounding box.

Sorghum Field Segmentation with U-Net from UAV RGB (무인기 기반 RGB 영상 활용 U-Net을 이용한 수수 재배지 분할)

  • Kisu Park;Chanseok Ryu ;Yeseong Kang;Eunri Kim;Jongchan Jeong;Jinki Park
    • Korean Journal of Remote Sensing
    • /
    • v.39 no.5_1
    • /
    • pp.521-535
    • /
    • 2023
  • When converting rice fields into fields,sorghum (sorghum bicolor L. Moench) has excellent moisture resistance, enabling stable production along with soybeans. Therefore, it is a crop that is expected to improve the self-sufficiency rate of domestic food crops and solve the rice supply-demand imbalance problem. However, there is a lack of fundamental statistics,such as cultivation fields required for estimating yields, due to the traditional survey method, which takes a long time even with a large manpower. In this study, U-Net was applied to RGB images based on unmanned aerial vehicle to confirm the possibility of non-destructive segmentation of sorghum cultivation fields. RGB images were acquired on July 28, August 13, and August 25, 2022. On each image acquisition date, datasets were divided into 6,000 training datasets and 1,000 validation datasets with a size of 512 × 512 images. Classification models were developed based on three classes consisting of Sorghum fields(sorghum), rice and soybean fields(others), and non-agricultural fields(background), and two classes consisting of sorghum and non-sorghum (others+background). The classification accuracy of sorghum cultivation fields was higher than 0.91 in the three class-based models at all acquisition dates, but learning confusion occurred in the other classes in the August dataset. In contrast, the two-class-based model showed an accuracy of 0.95 or better in all classes, with stable learning on the August dataset. As a result, two class-based models in August will be advantageous for calculating the cultivation fields of sorghum.

Detection of Gaze Direction for the Hearing-impaired in the Intelligent Space (지능형 공간에서 청각장애인의 시선 방향 검출)

  • Oh, Young-Joon;Hong, Kwang-Jin;Kim, Jong-In;Jung, Kee-Chul
    • The KIPS Transactions:PartB
    • /
    • v.18B no.6
    • /
    • pp.333-340
    • /
    • 2011
  • The Human-Computer Interaction(HCI) is a study of the method for interaction between human and computers that merges the ergonomics and the information technology. The intelligent space, which is a part of the HCI, is an important area to provide effective user interface for the disabled, who are alienated from the information-oriented society. In the intelligent space for the disabled, the method supporting information depends on types of disability. In this paper, we only support the hearing-impaired. It is material to the gaze direction detection method because it is very efficient information provide method to present information on gazing direction point, except for the information provide location perception method through directly contact with the hearing-impaired. We proposed the gaze direction detection method must be necessary in order to provide the residence life application to the hearing-impaired like this. The proposed method detects the region of the user from multi-view camera images, generates candidates for directions of gaze for horizontal and vertical from each camera, and calculates the gaze direction of the user through the comparison with the size of each candidate. In experimental results, the proposed method showed high detection rate with gaze direction and foot sensing rate with user's position, and showed the performance possibility of the scenario for the disabled.

A Fast Iris Region Finding Algorithm for Iris Recognition (홍채 인식을 위한 고속 홍채 영역 추출 방법)

  • 송선아;김백섭;송성호
    • Journal of KIISE:Software and Applications
    • /
    • v.30 no.9
    • /
    • pp.876-884
    • /
    • 2003
  • It is essential to identify both the pupil and iris boundaries for iris recognition. The circular edge detector proposed by Daugman is the most common and powerful method for the iris region extraction. The method is accurate but requires lots of computational time since it is based on the exhaustive search. Some heuristic methods have been proposed to reduce the computational time, but they are not as accurate as that of Daugman. In this paper, we propose a pupil and iris boundary finding algorithm which is faster than and as accurate as that of Daugman. The proposed algorithm searches the boundaries using the Daugman's circular edge detector, but reduces the search region using the problem domain knowledge. In order to find the pupil boundary, the search region is restricted in the maximum and minimum bounding circles in which the pupil resides. The bounding circles are obtained from the binarized pupil image. Two iris boundary points are obtained from the horizontal line passing through the center of the pupil region obtained above. These initial boundary points, together with the pupil point comprise two bounding circles. The iris boundary is searched in this bounding circles. Experiments show that the proposed algorithm is faster than that of Daugman and more accurate than the conventional heuristic methods.

Analysis of Land Cover Characteristics with Object-Based Classification Method - Focusing on the DMZ in Inje-gun, Gangwon-do - (객체기반 분류기법을 이용한 토지피복 특성분석 - 강원도 인제군의 DMZ지역 일원을 대상으로 -)

  • Na, Hyun-Sup;Lee, Jung-Soo
    • Journal of the Korean Association of Geographic Information Studies
    • /
    • v.17 no.2
    • /
    • pp.121-135
    • /
    • 2014
  • Object-based classification methods provide a valid alternative to traditional pixel-based methods. This study reports the results of an object-based classification to examine land cover in the demilitarized zones(DMZs) of Inje-gun. We used land cover classes(7 classes for main category and 13 classes for sub-category) selected from the criteria by Korea Ministry of Environment. The average and standard deviation of the spectrum values, and homogeneity of GLCM were chosen to map land cover types in an hierarchical approach using the nearest neighborhood method. We then identified the distributional characteristics of land cover by considering 3 topographic characteristics (altitude, slope gradient, distance from the Southern Limited Line(SLL)) within the DMZs. The results showed that scale 72, shape 0.2, color 0.8, compactness 0.5 and smoothness 0.5 were the optimum weight values while scale, shape and color were most influenced parameters in image segmentation. The forests (92%) were main land cover type in the DMZs; the grassland(5%), the urban area (2%) and the forests (broadleaf forest: 44%, mixed forest: 42%, coniferous forest: 6%) also occupied mostly in land cover classes for sub-category. The results also showed that facilities and roads had higher density within 2 km from the SLL, while paddy, field and bare land were distributed largely outside 6 km from the SLL. In addition, there was apparent distinction in land cover by topographic characteristics. The forest had higher density at above altitude 600m and above slope gradient $30^{\circ}$ while agriculture, bare land and grass land were distributed mainly at below altitude 600m and below slope gradient $30^{\circ}$.

Automatic Generation of Land Cover Map Using Residual U-Net (Residual U-Net을 이용한 토지피복지도 자동 제작 연구)

  • Yoo, Su Hong;Lee, Ji Sang;Bae, Jun Su;Sohn, Hong Gyoo
    • KSCE Journal of Civil and Environmental Engineering Research
    • /
    • v.40 no.5
    • /
    • pp.535-546
    • /
    • 2020
  • Land cover maps are derived from satellite and aerial images by the Ministry of Environment for the entire Korea since 1998. Even with their wide application in many sectors, their usage in research community is limited. The main reason for this is the map compilation cycle varies too much over the different regions. The situation requires us a new and quicker methodology for generating land cover maps. This study was conducted to automatically generate land cover map using aerial ortho-images and Landsat 8 satellite images. The input aerial and Landsat 8 image data were trained by Residual U-Net, one of the deep learning-based segmentation techniques. Study was carried out by dividing three groups. First and second group include part of level-II (medium) categories and third uses group level-III (large) classification category defined in land cover map. In the first group, the results using all 7 classes showed 86.6 % of classification accuracy The other two groups, which include level-II class, showed 71 % of classification accuracy. Based on the results of the study, the deep learning-based research for generating automatic level-III classification was presented.

Development of Velocity Imaging Method for Motility of Left Ventricle in Gated SPECT (게이트 심근 SPECT에서 좌심실의 운동성 분석을 위한 속도영상화 기법 개발)

  • Jo, Mi-Jung;Lee, Byeong-Il;Choi, Hyun-Ju;Hwang, Hae-Gil;Choi, Heung-Kook
    • Journal of Korea Multimedia Society
    • /
    • v.9 no.7
    • /
    • pp.808-817
    • /
    • 2006
  • Although the generally used the velocity index of doppler effect is a very significant factor in the functional evaluation of the left ventricle, it depends on the subjective evaluation of an inspector. The objective data of the motility can be obtained from the gated myocardial SPECT images by quantitative analysis. However, it is difficult to image visual of the velocity of the motion. The aim of our study is to develop a new method for the imaging velocity using the gated myocardial SPECT images and use it as an evaluation index for analyzing motility. First we visualized left ventricle into 3 dimensions using the coordinates of the points which were obtained through a segmentation of myocardium. Each point was represented by the different colors, according to the velocity of each point. We performed a validation study using 7 normal subjects and 15 myocardial infarction patients. To analyze motility, we used the average of the moved distance and the velocity. In normal cases, the average of the moved distance was 4.3mm and the average of the velocity was 11.9mm. In patient cases, the average of the moved distance was 3.9mm and the average of the velocity was 10.5mm. These results show that the motility of normal subjects is higher than the abnormal subjects. We expect that our proposed method could become a way to improve the accuracy and reproducibility for the functional evaluation of myocardial wall.

  • PDF

Automatic Extraction of Ascending Aorta and Ostium in Cardiac CT Angiography Images (심장 CT 혈관 조영 영상에서 대동맥 및 심문 자동 검출)

  • Kim, Hye-Ryun;Kang, Mi-Sun;Kim, Myoung-Hee
    • Journal of the Korea Computer Graphics Society
    • /
    • v.23 no.1
    • /
    • pp.49-55
    • /
    • 2017
  • Computed tomographic angiography (CTA) is widely used in the diagnosis and treatment of coronary artery disease because it shows not only the whole anatomical structure of the cardiovascular three-dimensionally but also provides information on the lesion and type of plaque. However, due to the large size of the image, there is a limitation in manually extracting coronary arteries, and related researches are performed to automatically extract coronary arteries accurately. As the coronary artery originate from the ascending aorta, the ascending aorta and ostium should be detected to extract the coronary tree accurately. In this paper, we propose an automatic segmentation for the ostium as a starting structure of coronary artery in CTA. First, the region of the ascending aorta is initially detected by using Hough circle transform based on the relative position and size of the ascending aorta. Second, the volume of interest is defined to reduce the search range based on the initial area. Third, the refined ascending aorta is segmented by using a two-dimensional geodesic active contour. Finally, the two ostia are detected within the region of the refined ascending aorta. For the evaluation of our method, we measured the Euclidean distance between the result and the ground truths annotated manually by medical experts in 20 CTA images. The experimental results showed that the ostia were accurately detected.

Reversible Watermarking based on Predicted Error Histogram for Medical Imagery (의료 영상을 위한 추정오차 히스토그램 기반 가역 워터마킹 알고리즘)

  • Oh, Gi-Tae;Jang, Han-Byul;Do, Um-Ji;Lee, Hae-Yeoun
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.4 no.5
    • /
    • pp.231-240
    • /
    • 2015
  • Medical imagery require to protect the privacy with preserving the quality of the original contents. Therefore, reversible watermarking is a solution for this purpose. Previous researches have focused on general imagery and achieved high capacity and high quality. However, they raise a distortion over entire image and hence are not applicable to medical imagery which require to preserve the quality of the objects. In this paper, we propose a novel reversible watermarking for medical imagery, which preserve the quality of the objects and achieves high capacity. First, object and background region is segmented and then predicted error histogram-based reversible watermarking is applied for each region. For the efficient watermark embedding with small distortion in the object region, the embedding level at object region is set as low while the embedding level at background region is set as high. In experiments, the proposed algorithm is compared with the previous predicted error histogram-based algorithm in aspects of embedding capacity and perceptual quality. Results support that the proposed algorithm performs well over the previous algorithm.