• Title/Summary/Keyword: SegNet

Search Result 22, Processing Time 0.03 seconds

Extracting Flooded Areas in Southeast Asia Using SegNet and U-Net (SegNet과 U-Net을 활용한 동남아시아 지역 홍수탐지)

  • Kim, Junwoo;Jeon, Hyungyun;Kim, Duk-jin
    • Korean Journal of Remote Sensing
    • /
    • v.36 no.5_3
    • /
    • pp.1095-1107
    • /
    • 2020
  • Flood monitoring using satellite data has been constrained by obtaining satellite images for flood peak and accurately extracting flooded areas from satellite data. Deep learning is a promising method for satellite image classification, yet the potential of deep learning-based flooded area extraction using SAR data remained uncertain, which has advantages in obtaining data, comparing to optical satellite data. This research explores the performance of SegNet and U-Net on image segmentation by extracting flooded areas in the Khorat basin, Mekong river basin, and Cagayan river basin in Thailand, Laos, and the Philippines from Sentinel-1 A/B satellite data. Results show that Global Accuracy, Mean IoU, and Mean BF Score of SegNet are 0.9847, 0.6016, and 0.6467 respectively, whereas those of U-Net are 0.9937, 0.7022, 0.7125. Visual interpretation shows that the classification accuracy of U-Net is higher than SegNet, but overall processing time of SegNet is around three times faster than that of U-Net. It is anticipated that the results of this research could be used when developing deep learning-based flood monitoring models and presenting fully automated flooded area extraction models.

Evaluation on the Usefulness of X-ray Computer-Aided Detection (CAD) System for Pulmonary Tuberculosis (PTB) using SegNet (X-ray 영상에서 SegNet을 이용한 폐결핵 자동검출 시스템의 유용성 평가)

  • Lee, J.H.;Ahn, H.S.;Choi, D.H.;Tae, Ki Sik
    • Journal of Biomedical Engineering Research
    • /
    • v.38 no.1
    • /
    • pp.25-31
    • /
    • 2017
  • Testing TB in chest X-ray images is a typical method to diagnose presence and magnitude of PTB lesion. However, the method has limitation due to inter-reader variability. Therefore, it is essential to overcome this drawback with automatic interpretation. In this study, we propose a novel method for detection of PTB using SegNet, which is a deep learning architecture for semantic pixel wise image labelling. SegNet is composed of a stack of encoders followed by a corresponding decoder stack which feeds into a soft-max classification layer. We modified parameters of SegNet to change the number of classes from 12 to 2 (TB or none-TB) and applied the architecture to automatically interpret chest radiographs. 552 chest X-ray images, provided by The Korean Institute of Tuberculosis, used for training and test and we constructed a receiver operating characteristic (ROC) curve. As a consequence, the area under the curve (AUC) was 90.4% (95% CI:[85.1, 95.7]) with a classification accuracy of 84.3%. A sensitivity was 85.7% and specificity was 82.8% on 431 training images (TB 172, none-TB 259) and 121 test images (TB 63, none-TB 58). This results show that detecting PTB using SegNet is comparable to other PTB detection methods.

Land Cover Classification Using Sematic Image Segmentation with Deep Learning (딥러닝 기반의 영상분할을 이용한 토지피복분류)

  • Lee, Seonghyeok;Kim, Jinsoo
    • Korean Journal of Remote Sensing
    • /
    • v.35 no.2
    • /
    • pp.279-288
    • /
    • 2019
  • We evaluated the land cover classification performance of SegNet, which features semantic segmentation of aerial imagery. We selected four semantic classes, i.e., urban, farmland, forest, and water areas, and created 2,000 datasets using aerial images and land cover maps. The datasets were divided at a 8:2 ratio into training (1,600) and validation datasets (400); we evaluated validation accuracy after tuning the hyperparameters. SegNet performance was optimal at a batch size of five with 100,000 iterations. When 200 test datasets were subjected to semantic segmentation using the trained SegNet model, the accuracies were farmland 87.89%, forest 87.18%, water 83.66%, and urban regions 82.67%; the overall accuracy was 85.48%. Thus, deep learning-based semantic segmentation can be used to classify land cover.

Detection of Zebra-crossing Areas Based on Deep Learning with Combination of SegNet and ResNet (SegNet과 ResNet을 조합한 딥러닝에 기반한 횡단보도 영역 검출)

  • Liang, Han;Seo, Suyoung
    • Journal of the Korean Society of Surveying, Geodesy, Photogrammetry and Cartography
    • /
    • v.39 no.3
    • /
    • pp.141-148
    • /
    • 2021
  • This paper presents a method to detect zebra-crossing using deep learning which combines SegNet and ResNet. For the blind, a safe crossing system is important to know exactly where the zebra-crossings are. Zebra-crossing detection by deep learning can be a good solution to this problem and robotic vision-based assistive technologies sprung up over the past few years, which focused on specific scene objects using monocular detectors. These traditional methods have achieved significant results with relatively long processing times, and enhanced the zebra-crossing perception to a large extent. However, running all detectors jointly incurs a long latency and becomes computationally prohibitive on wearable embedded systems. In this paper, we propose a model for fast and stable segmentation of zebra-crossing from captured images. The model is improved based on a combination of SegNet and ResNet and consists of three steps. First, the input image is subsampled to extract image features and the convolutional neural network of ResNet is modified to make it the new encoder. Second, through the SegNet original up-sampling network, the abstract features are restored to the original image size. Finally, the method classifies all pixels and calculates the accuracy of each pixel. The experimental results prove the efficiency of the modified semantic segmentation algorithm with a relatively high computing speed.

The Comparison of Segmentation Performance between SegFormer and U-Net on Railway Components (SegFormer 및 U-Net의 철도 구성요소 객체 분할 성능 비교)

  • Jaehyun Lee;Changjoon Park;Namjung Kim;Junhwi Park;Jeonghwan Gwak
    • Proceedings of the Korean Society of Computer Information Conference
    • /
    • 2024.01a
    • /
    • pp.347-348
    • /
    • 2024
  • 본 논문에서는 철도 구성요소 모니터링을 위한 효율적인 객체 분할 기법으로 사전학습된 SegFormer 모델의 적용을 제안하고, 객체 분할을 위해 보편적으로 사용되는 U-Net 모델과의 성능 비교 분석을 진행하였다. 철도의 주요 구성요소인 선로, 침목, 고정 장치, 배경을 분할할 수 있도록 라벨링된 데이터셋을 학습에 사용하였다. SegFormer 모델이 대조군인 U-Net보다 성능이 Jaccard Score 기준 5.29% 향상됨에 따라 Vision Transformer 기반의 모델이 기존 CNN 기반 모델의 이미지의 전역적인 문맥을 파악하기 상대적으로 어렵다는 한계를 극복하고, 철도 구성요소 객체 분할에 더욱 효율적인 모델임을 확인한다.

  • PDF

Multi-scale U-SegNet architecture with cascaded dilated convolutions for brain MRI Segmentation

  • Dayananda, Chaitra;Lee, Bumshik
    • Proceedings of the Korean Society of Broadcast Engineers Conference
    • /
    • 2020.11a
    • /
    • pp.25-28
    • /
    • 2020
  • Automatic segmentation of brain tissues such as WM, GM, and CSF from brain MRI scans is helpful for the diagnosis of many neurological disorders. Accurate segmentation of these brain structures is a very challenging task due to low tissue contrast, bias filed, and partial volume effects. With the aim to improve brain MRI segmentation accuracy, we propose an end-to-end convolutional based U-SegNet architecture designed with multi-scale kernels, which includes cascaded dilated convolutions for the task of brain MRI segmentation. The multi-scale convolution kernels are designed to extract abundant semantic features and capture context information at different scales. Further, the cascaded dilated convolution scheme helps to alleviate the vanishing gradient problem in the proposed model. Experimental outcomes indicate that the proposed architecture is superior to the traditional deep-learning methods such as Segnet, U-net, and U-Segnet and achieves high performance with an average DSC of 93% and 86% of JI value for brain MRI segmentation.

  • PDF

Enhanced CNN Model for Brain Tumor Classification

  • Kasukurthi, Aravinda;Paleti, Lakshmikanth;Brahmaiah, Madamanchi;Sree, Ch.Sudha
    • International Journal of Computer Science & Network Security
    • /
    • v.22 no.5
    • /
    • pp.143-148
    • /
    • 2022
  • Brain tumor classification is an important process that allows doctors to plan treatment for patients based on the stages of the tumor. To improve classification performance, various CNN-based architectures are used for brain tumor classification. Existing methods for brain tumor segmentation suffer from overfitting and poor efficiency when dealing with large datasets. The enhanced CNN architecture proposed in this study is based on U-Net for brain tumor segmentation, RefineNet for pattern analysis, and SegNet architecture for brain tumor classification. The brain tumor benchmark dataset was used to evaluate the enhanced CNN model's efficiency. Based on the local and context information of the MRI image, the U-Net provides good segmentation. SegNet selects the most important features for classification while also reducing the trainable parameters. In the classification of brain tumors, the enhanced CNN method outperforms the existing methods. The enhanced CNN model has an accuracy of 96.85 percent, while the existing CNN with transfer learning has an accuracy of 94.82 percent.

Comparison of Performance of Medical Image Semantic Segmentation Model in ATLASV2.0 Data (ATLAS V2.0 데이터에서 의료영상 분할 모델 성능 비교)

  • So Yeon Woo;Yeong Hyeon Gu;Seong Joon Yoo
    • Journal of Broadcast Engineering
    • /
    • v.28 no.3
    • /
    • pp.267-274
    • /
    • 2023
  • There is a problem that the size of the dataset is insufficient due to the limitation of the collection of the medical image public data, so there is a possibility that the existing studies are overfitted to the public dataset. In this paper, we compare the performance of eight (Unet, X-Net, HarDNet, SegNet, PSPNet, SwinUnet, 3D-ResU-Net, UNETR) medical image semantic segmentation models to revalidate the superiority of existing models. Anatomical Tracings of Lesions After Stroke (ATLAS) V1.2, a public dataset for stroke diagnosis, is used to compare the performance of the models and the performance of the models in ATLAS V2.0. Experimental results show that most models have similar performance in V1.2 and V2.0, but X-net and 3D-ResU-Net have higher performance in V1.2 datasets. These results can be interpreted that the models may be overfitted to V1.2.

High-performance of Deep learning Colorization With Wavelet fusion (웨이블릿 퓨전에 의한 딥러닝 색상화의 성능 향상)

  • Kim, Young-Back;Choi, Hyun;Cho, Joong-Hwee
    • IEMEK Journal of Embedded Systems and Applications
    • /
    • v.13 no.6
    • /
    • pp.313-319
    • /
    • 2018
  • We propose a post-processing algorithm to improve the quality of the RGB image generated by deep learning based colorization from the gray-scale image of an infrared camera. Wavelet fusion is used to generate a new luminance component of the RGB image luminance component from the deep learning model and the luminance component of the infrared camera. PSNR is increased for all experimental images by applying the proposed algorithm to RGB images generated by two deep learning models of SegNet and DCGAN. For the SegNet model, the average PSNR is improved by 1.3906dB at level 1 of the Haar wavelet method. For the DCGAN model, PSNR is improved 0.0759dB on the average at level 5 of the Daubechies wavelet method. It is also confirmed that the edge components are emphasized by the post-processing and the visibility is improved.