• Title/Summary/Keyword: Deep learning segmentation

Search Result 379, Processing Time 0.025 seconds

Atrous Residual U-Net for Semantic Segmentation in Street Scenes based on Deep Learning (딥러닝 기반 거리 영상의 Semantic Segmentation을 위한 Atrous Residual U-Net)

  • Shin, SeokYong;Lee, SangHun;Han, HyunHo
    • Journal of Convergence for Information Technology
    • /
    • v.11 no.10
    • /
    • pp.45-52
    • /
    • 2021
  • In this paper, we proposed an Atrous Residual U-Net (AR-UNet) to improve the segmentation accuracy of semantic segmentation method based on U-Net. The U-Net is mainly used in fields such as medical image analysis, autonomous vehicles, and remote sensing images. The conventional U-Net lacks extracted features due to the small number of convolution layers in the encoder part. The extracted features are essential for classifying object categories, and if they are insufficient, it causes a problem of lowering the segmentation accuracy. Therefore, to improve this problem, we proposed the AR-UNet using residual learning and ASPP in the encoder. Residual learning improves feature extraction ability and is effective in preventing feature loss and vanishing gradient problems caused by continuous convolutions. In addition, ASPP enables additional feature extraction without reducing the resolution of the feature map. Experiments verified the effectiveness of the AR-UNet with Cityscapes dataset. The experimental results showed that the AR-UNet showed improved segmentation results compared to the conventional U-Net. In this way, AR-UNet can contribute to the advancement of many applications where accuracy is important.

Deep Learning-based Rice Seed Segmentation for Phynotyping (표현체 연구를 위한 심화학습 기반 벼 종자 분할)

  • Jeong, Yu Seok;Lee, Hong Ro;Baek, Jeong Ho;Kim, Kyung Hwan;Chung, Young Suk;Lee, Chang Woo
    • Journal of Korea Society of Industrial Information Systems
    • /
    • v.25 no.5
    • /
    • pp.23-29
    • /
    • 2020
  • The National Institute of Agricultural Sciences of the Rural Developement Administration (NAS, RDA) is conducting various studies on various crops, such as monitoring the cultivation environment and analyzing harvested seeds for high-throughput phenotyping. In this paper, we propose a deep learning-based rice seed segmentation method to analyze the seeds of various crops owned by the NAS. Using Mask-RCNN deep learning model, we perform the rice seed segmentation from manually taken images under specific environment (constant lighting, white background) for analyzing the seed characteristics. For this purpose, we perform the parameter tuning process of the Mask-RCNN model. By the proposed method, the results of the test on seed object detection showed that the accuracy was 82% for rice stem image and 97% for rice grain image, respectively. As a future study, we are planning to researches of more reliable seeds extraction from cluttered seed images by a deep learning-based approach and selection of high-throughput phenotype through precise data analysis such as length, width, and thickness from the detected seed objects.

Segmentation of underwater images using morphology for deep learning (딥러닝을 위한 모폴로지를 이용한 수중 영상의 세그먼테이션)

  • Ji-Eun Lee;Chul-Won Lee;Seok-Joon Park;Jea-Beom Shin;Hyun-Gi Jung
    • The Journal of the Acoustical Society of Korea
    • /
    • v.42 no.4
    • /
    • pp.370-376
    • /
    • 2023
  • In the underwater image, it is not clear to distinguish the shape of the target due to underwater noise and low resolution. In addition, as an input of deep learning, underwater images require pre-processing and segmentation must be preceded. Even after pre-processing, the target is not clear, and the performance of detection and identification by deep learning may not be high. Therefore, it is necessary to distinguish and clarify the target. In this study, the importance of target shadows is confirmed in underwater images, object detection and target area acquisition by shadows, and data containing only the shape of targets and shadows without underwater background are generated. We present the process of converting the shadow image into a 3-mode image in which the target is white, the shadow is black, and the background is gray. Through this, it is possible to provide an image that is clearly pre-processed and easily discriminated as an input of deep learning. In addition, if the image processing code using Open Source Computer Vision (OpenCV)Library was used for processing, the processing speed was also suitable for real-time processing.

An Automatic Breast Mass Segmentation based on Deep Learning on Mammogram (유방 영상에서 딥러닝 기반의 유방 종괴 자동 분할 연구)

  • Kwon, So Yoon;Kim, Young Jae;Kim, Gwang Gi
    • Journal of Korea Multimedia Society
    • /
    • v.21 no.12
    • /
    • pp.1363-1369
    • /
    • 2018
  • Breast cancer is one of the most common cancers in women worldwide. In Korea, breast cancer is most common cancer in women followed by thyroid cancer. The purpose of this study is to evaluate the possibility of using deep - run model for segmentation of breast masses and to identify the best deep-run model for breast mass segmentation. In this study, data of patients with breast masses were collected at Asan Medical Center. We used 596 images of mammography and 596 images of gold standard. In the area of interest of the medical image, it was cut into a rectangular shape with a margin of about 10% up and down, and then converted into an 8-bit image by adjusting the window width and level. Also, the size of the image was resampled to $150{\times}150$. In Deconvolution net, the average accuracy is 91.78%. In U-net, the average accuracy is 90.09%. Deconvolution net showed slightly better performance than U-net in this study, so it is expected that deconvolution net will be better for breast mass segmentation. However, because of few cases, there are a few images that are not accurately segmented. Therefore, more research is needed with various training data.

Convolutional Neural Network-Based Automatic Segmentation of Substantia Nigra on Nigrosome and Neuromelanin Sensitive MR Images

  • Kang, Junghwa;Kim, Hyeonha;Kim, Eunjin;Kim, Eunbi;Lee, Hyebin;Shin, Na-young;Nam, Yoonho
    • Investigative Magnetic Resonance Imaging
    • /
    • v.25 no.3
    • /
    • pp.156-163
    • /
    • 2021
  • Recently, neuromelanin and nigrosome imaging techniques have been developed to evaluate the substantia nigra in Parkinson's disease. Previous studies have shown potential benefits of quantitative analysis of neuromelanin and nigrosome images in the substantia nigra, although visual assessments have been performed to evaluate structures in most studies. In this study, we investigate the potential of using deep learning based automatic region segmentation techniques for quantitative analysis of the substantia nigra. The deep convolutional neural network was trained to automatically segment substantia nigra regions on 3D nigrosome and neuromelanin sensitive MR images obtained from 30 subjects. With a 5-fold cross-validation, the mean calculated dice similarity coefficient between manual and deep learning was 0.70 ± 0.11. Although calculated dice similarity coefficients were relatively low due to empirically drawn margins, selected slices were overlapped for more than two slices of all subjects. Our results demonstrate that deep convolutional neural network-based method could provide reliable localization of substantia nigra regions on neuromelanin and nigrosome sensitive MR images.

An evaluation methodology for cement concrete lining crack segmentation deep learning model (콘크리트 라이닝 균열 분할 딥러닝 모델 평가 방법)

  • Ham, Sangwoo;Bae, Soohyeon;Lee, Impyeong;Lee, Gyu-Phil;Kim, Donggyou
    • Journal of Korean Tunnelling and Underground Space Association
    • /
    • v.24 no.6
    • /
    • pp.513-524
    • /
    • 2022
  • Recently, detecting damages of civil infrastructures from digital images using deep learning technology became a very popular research topic. In order to adapt those methodologies to the field, it is essential to explain robustness of deep learning models. Our research points out that the existing pixel-based deep learning model evaluation metrics are not sufficient for detecting cracks since cracks have linear appearance, and proposes a new evaluation methodology to explain crack segmentation deep learning model more rationally. Specifically, we design, implement and validate a methodology to generate tolerance buffer alongside skeletonized ground truth data and prediction results to consider overall similarity of topology of the ground truth and the prediction rather than pixel-wise accuracy. We could overcome over-estimation or under-estimation problem of crack segmentation model evaluation through using our methodology, and we expect that our methodology can explain crack segmentation deep learning models better.

Development of Virtual Simulator and Database for Deep Learning-based Object Detection (딥러닝 기반 장애물 인식을 위한 가상환경 및 데이터베이스 구축)

  • Lee, JaeIn;Gwak, Gisung;Kim, KyongSu;Kang, WonYul;Shin, DaeYoung;Hwang, Sung-Ho
    • Journal of Drive and Control
    • /
    • v.18 no.4
    • /
    • pp.9-18
    • /
    • 2021
  • This study proposes a method for creating learning datasets to recognize obstacles using deep learning algorithms in automated construction machinery or an autonomous vehicle. Recently, many researchers and engineers have developed various recognition algorithms based on deep learning following an increase in computing power. In particular, the image classification technology and image segmentation technology represent deep learning recognition algorithms. They are used to identify obstacles that interfere with the driving situation of an autonomous vehicle. Therefore, various organizations and companies have started distributing open datasets, but there is a remote possibility that they will perfectly match the user's desired environment. In this study, we created an interface of the virtual simulator such that users can easily create their desired training dataset. In addition, the customized dataset was further advanced by using the RDBMS system, and the recognition rate was improved.

A Study of Real-time Semantic Segmentation Performance Improvement in Unstructured Outdoor Environment (비정형 야지환경 주행상황에서의 실시간 의미론적 영상 분할 알고리즘 성능 향상에 관한 연구)

  • Daeyoung, Kim;Seunguk, Ahn;Seung-Woo, Seo
    • Journal of the Korea Institute of Military Science and Technology
    • /
    • v.25 no.6
    • /
    • pp.606-616
    • /
    • 2022
  • Semantic segmentation in autonomous driving for unstructured environments is challenging due to the presence of uneven terrains, unstructured class boundaries, irregular features and strong textures. Current off-road datasets exhibit difficulties like class imbalance and understanding of varying environmental topography. To overcome these issues, we propose a deep learning framework for semantic segmentation that involves a pooled class semantic segmentation with five classes. The evaluation of the framework is carried out on two off-road driving datasets, RUGD and TAS500. The results show that our proposed method achieves high accuracy and real-time performance.

A review of Chinese named entity recognition

  • Cheng, Jieren;Liu, Jingxin;Xu, Xinbin;Xia, Dongwan;Liu, Le;Sheng, Victor S.
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.15 no.6
    • /
    • pp.2012-2030
    • /
    • 2021
  • Named Entity Recognition (NER) is used to identify entity nouns in the corpus such as Location, Person and Organization, etc. NER is also an important basic of research in various natural language fields. The processing of Chinese NER has some unique difficulties, for example, there is no obvious segmentation boundary between each Chinese character in a Chinese sentence. The Chinese NER task is often combined with Chinese word segmentation, and so on. In response to these problems, we summarize the recognition methods of Chinese NER. In this review, we first introduce the sequence labeling system and evaluation metrics of NER. Then, we divide Chinese NER methods into rule-based methods, statistics-based machine learning methods and deep learning-based methods. Subsequently, we analyze in detail the model framework based on deep learning and the typical Chinese NER methods. Finally, we put forward the current challenges and future research directions of Chinese NER technology.

Accuracy evaluation of liver and tumor auto-segmentation in CT images using 2D CoordConv DeepLab V3+ model in radiotherapy

  • An, Na young;Kang, Young-nam
    • Journal of Biomedical Engineering Research
    • /
    • v.43 no.5
    • /
    • pp.341-352
    • /
    • 2022
  • Medical image segmentation is the most important task in radiation therapy. Especially, when segmenting medical images, the liver is one of the most difficult organs to segment because it has various shapes and is close to other organs. Therefore, automatic segmentation of the liver in computed tomography (CT) images is a difficult task. Since tumors also have low contrast in surrounding tissues, and the shape, location, size, and number of tumors vary from patient to patient, accurate tumor segmentation takes a long time. In this study, we propose a method algorithm for automatically segmenting the liver and tumor for this purpose. As an advantage of setting the boundaries of the tumor, the liver and tumor were automatically segmented from the CT image using the 2D CoordConv DeepLab V3+ model using the CoordConv layer. For tumors, only cropped liver images were used to improve accuracy. Additionally, to increase the segmentation accuracy, augmentation, preprocess, loss function, and hyperparameter were used to find optimal values. We compared the CoordConv DeepLab v3+ model using the CoordConv layer and the DeepLab V3+ model without the CoordConv layer to determine whether they affected the segmentation accuracy. The data sets used included 131 hepatic tumor segmentation (LiTS) challenge data sets (100 train sets, 16 validation sets, and 15 test sets). Additional learned data were tested using 15 clinical data from Seoul St. Mary's Hospital. The evaluation was compared with the study results learned with a two-dimensional deep learning-based model. Dice values without the CoordConv layer achieved 0.965 ± 0.01 for liver segmentation and 0.925 ± 0.04 for tumor segmentation using the LiTS data set. Results from the clinical data set achieved 0.927 ± 0.02 for liver division and 0.903 ± 0.05 for tumor division. The dice values using the CoordConv layer achieved 0.989 ± 0.02 for liver segmentation and 0.937 ± 0.07 for tumor segmentation using the LiTS data set. Results from the clinical data set achieved 0.944 ± 0.02 for liver division and 0.916 ± 0.18 for tumor division. The use of CoordConv layers improves the segmentation accuracy. The highest of the most recently published values were 0.960 and 0.749 for liver and tumor division, respectively. However, better performance was achieved with 0.989 and 0.937 results for liver and tumor, which would have been used with the algorithm proposed in this study. The algorithm proposed in this study can play a useful role in treatment planning by improving contouring accuracy and reducing time when segmentation evaluation of liver and tumor is performed. And accurate identification of liver anatomy in medical imaging applications, such as surgical planning, as well as radiotherapy, which can leverage the findings of this study, can help clinical evaluation of the risks and benefits of liver intervention.