• 제목/요약/키워드: Image Learning

검색결과 3,146건 처리시간 0.031초

분해 심층 학습을 이용한 저조도 영상 개선 방식 (Low-light Image Enhancement Method Using Decomposition-based Deep-Learning)

  • 오종근;홍민철
    • 전기전자학회논문지
    • /
    • 제25권1호
    • /
    • pp.139-147
    • /
    • 2021
  • 본 논문에서는 저조도 영상을 개선하기 위한 영상 분해 기반 심층 학습 방법 및 분해 채널 특성에 따른 손실함수를 제안한다. 기존 기법들의 문제점인 색신호 왜곡 및 할로 현상을 제거하기 위해, 입력 영상의 휘도 채널을 반사 성분과 조도 성분으로 분해하고, 반사 성분, 조도 성분 및 색차 신호를 신호 특성에 적합한 심층학습 과정을 적용하는 분해 기반 다중 구조 심층 학습 방법을 제안한다. 더불어, 분해 채널들의 특성에 따른 혼합 놈 기반의 손실함수를 정의하여 복원 영상의 안정성을 증대하고 열화 현상을 제거하기 위한 기법에 대해 기술한다. 실험 결과를 통해 제안한 방법이 다양한 저조도 영상을 효과적으로 개선하였음을 확인할 수 있었다.

내용 기반 이미지 검색을 위한 개선된 SIM 방법 (Improved SIM Algorithm for Contents-based Image Retrieval)

  • 김광백
    • 지능정보연구
    • /
    • 제15권2호
    • /
    • pp.49-59
    • /
    • 2009
  • 내용기반 이미지 검색은 색상, 질감 등의 이미지 자체의 자질들을 이용하여 검색하므로 텍스트 기반 이미지 검색의 객관성 부족과 모든 이미지에 사람이 주석을 달아야 하는 단점을 보완할 수 있는 이미지 검색 방법이다. 이러한 내용 기반 이미지 검색에서 사용되는 방식 중 SIM(Self-organizing Image browsing Map) 방식은 SOM 알고리즘을 이용하여 이미지들을 브라우징 가능한 그룹으로 맵핑하고 그 결과를 바탕으로 이미지를 검색하게 된다. 하지만 비슷한 이미지라 할지라도 이미지의 밝기, 피사체의 움직임 등에 의하여 색상 정보가 다르게 나타나게 되면 SOM 알고리즘의 학습 과정에서 유사한 이미지들을 그룹화한 노드를 BMU로 선택하지 못하고 떨어져 있는 다른 노드를 선택하게 된다. 이 경우 학습이 진행되면서 유사한 이미지들이 군집하는 과정을 거치지만 학습이 완료될 때까지 다른 유사 이미지들을 그룹화한 노드에 맵핑이 되지 못하는 경우가 발생한다. 그 결과, 검색 결과에 나타나지 못하여 적합 이미지 검색률이 낮아 질 수 있다. 따라서 본 논문에서는 HSV 색상모델을 이용하여 양자화하고 이미지의 색상 특징 벡터를 추출한 뒤 SOM 알고리즘을 이용하여 이미지들을 브라우징 가능한 그룹으로 맵핑한다. 이때 SIM 방식의 문제점인 유사 이미지가 따로 맵핑되어 적합 이미지 검색률이 낮아지는 것을 줄이기 위하여 SOM을 두 개의 층으로 구성한다. 첫 번째 층에서 이미지의 색상 자질을 이용하여 학습을 완료한 후, 학습이 완료된 첫 번째 층 맵의 각 노드들의 연결 가중치를 이용하여 두 번째 층에서 다시 한번 학습을 수행한다. 두 개의 층으로 학습이 완료된 두 번째 층의 SOM에 질의 이미지의 특징 벡터를 입력하여 BMU를 선택하고 BMU와 연결된 첫 번째 층의 노드를 최종 선택하여 이미지를 검색한다. 실험결과, 제안된 이미지 검색 방법이 기존의 이미지 검색 방법 보다 적합 이미지의 검색 성공률이 높은 것을 확인 할 수 있었다.

  • PDF

Deep Learning-Based Computed Tomography Image Standardization to Improve Generalizability of Deep Learning-Based Hepatic Segmentation

  • Seul Bi Lee;Youngtaek Hong;Yeon Jin Cho;Dawun Jeong;Jina Lee;Soon Ho Yoon;Seunghyun Lee;Young Hun Choi;Jung-Eun Cheon
    • Korean Journal of Radiology
    • /
    • 제24권4호
    • /
    • pp.294-304
    • /
    • 2023
  • Objective: We aimed to investigate whether image standardization using deep learning-based computed tomography (CT) image conversion would improve the performance of deep learning-based automated hepatic segmentation across various reconstruction methods. Materials and Methods: We collected contrast-enhanced dual-energy CT of the abdomen that was obtained using various reconstruction methods, including filtered back projection, iterative reconstruction, optimum contrast, and monoenergetic images with 40, 60, and 80 keV. A deep learning based image conversion algorithm was developed to standardize the CT images using 142 CT examinations (128 for training and 14 for tuning). A separate set of 43 CT examinations from 42 patients (mean age, 10.1 years) was used as the test data. A commercial software program (MEDIP PRO v2.0.0.0, MEDICALIP Co. Ltd.) based on 2D U-NET was used to create liver segmentation masks with liver volume. The original 80 keV images were used as the ground truth. We used the paired t-test to compare the segmentation performance in the Dice similarity coefficient (DSC) and difference ratio of the liver volume relative to the ground truth volume before and after image standardization. The concordance correlation coefficient (CCC) was used to assess the agreement between the segmented liver volume and ground-truth volume. Results: The original CT images showed variable and poor segmentation performances. The standardized images achieved significantly higher DSCs for liver segmentation than the original images (DSC [original, 5.40%-91.27%] vs. [standardized, 93.16%-96.74%], all P < 0.001). The difference ratio of liver volume also decreased significantly after image conversion (original, 9.84%-91.37% vs. standardized, 1.99%-4.41%). In all protocols, CCCs improved after image conversion (original, -0.006-0.964 vs. standardized, 0.990-0.998). Conclusion: Deep learning-based CT image standardization can improve the performance of automated hepatic segmentation using CT images reconstructed using various methods. Deep learning-based CT image conversion may have the potential to improve the generalizability of the segmentation network.

Additional Learning Framework for Multipurpose Image Recognition

  • Itani, Michiaki;Iyatomi, Hitoshi;Hagiwara, Masafumi
    • 한국지능시스템학회:학술대회논문집
    • /
    • 한국퍼지및지능시스템학회 2003년도 ISIS 2003
    • /
    • pp.480-483
    • /
    • 2003
  • We propose a new framework that aims at multi-purpose image recognition, a difficult task for the conventional rule-based systems. This framework is farmed based on the idea of computer-based learning algorithm. In this research, we introduce the new functions of an additional learning and a knowledge reconstruction on the Fuzzy Inference Neural Network (FINN) (1) to enable the system to accommodate new objects and enhance the accuracy as necessary. We examine the capability of the proposed framework using two examples. The first one is the capital letter recognition task from UCI machine learning repository to estimate the effectiveness of the framework itself, Even though the whole training data was not given in advance, the proposed framework operated with a small loss of accuracy by introducing functions of the additional learning and the knowledge reconstruction. The other is the scenery image recognition. We confirmed that the proposed framework could recognize images with high accuracy and accommodate new object recursively.

  • PDF

A Deep Learning Approach for Classification of Cloud Image Patches on Small Datasets

  • Phung, Van Hiep;Rhee, Eun Joo
    • Journal of information and communication convergence engineering
    • /
    • 제16권3호
    • /
    • pp.173-178
    • /
    • 2018
  • Accurate classification of cloud images is a challenging task. Almost all the existing methods rely on hand-crafted feature extraction. Their limitation is low discriminative power. In the recent years, deep learning with convolution neural networks (CNNs), which can auto extract features, has achieved promising results in many computer vision and image understanding fields. However, deep learning approaches usually need large datasets. This paper proposes a deep learning approach for classification of cloud image patches on small datasets. First, we design a suitable deep learning model for small datasets using a CNN, and then we apply data augmentation and dropout regularization techniques to increase the generalization of the model. The experiments for the proposed approach were performed on SWIMCAT small dataset with k-fold cross-validation. The experimental results demonstrated perfect classification accuracy for most classes on every fold, and confirmed both the high accuracy and the robustness of the proposed model.

웨이블릿 퓨전에 의한 딥러닝 색상화의 성능 향상 (High-performance of Deep learning Colorization With Wavelet fusion)

  • 김영백;최현;조중휘
    • 대한임베디드공학회논문지
    • /
    • 제13권6호
    • /
    • pp.313-319
    • /
    • 2018
  • We propose a post-processing algorithm to improve the quality of the RGB image generated by deep learning based colorization from the gray-scale image of an infrared camera. Wavelet fusion is used to generate a new luminance component of the RGB image luminance component from the deep learning model and the luminance component of the infrared camera. PSNR is increased for all experimental images by applying the proposed algorithm to RGB images generated by two deep learning models of SegNet and DCGAN. For the SegNet model, the average PSNR is improved by 1.3906dB at level 1 of the Haar wavelet method. For the DCGAN model, PSNR is improved 0.0759dB on the average at level 5 of the Daubechies wavelet method. It is also confirmed that the edge components are emphasized by the post-processing and the visibility is improved.

Deep Learning Machine Vision System with High Object Recognition Rate using Multiple-Exposure Image Sensing Method

  • Park, Min-Jun;Kim, Hyeon-June
    • 센서학회지
    • /
    • 제30권2호
    • /
    • pp.76-81
    • /
    • 2021
  • In this study, we propose a machine vision system with a high object recognition rate. By utilizing a multiple-exposure image sensing technique, the proposed deep learning-based machine vision system can cover a wide light intensity range without further learning processes on the various light intensity range. If the proposed machine vision system fails to recognize object features, the system operates in a multiple-exposure sensing mode and detects the target object that is blocked in the near dark or bright region. Furthermore, short- and long-exposure images from the multiple-exposure sensing mode are synthesized to obtain accurate object feature information. That results in the generation of a wide dynamic range of image information. Even with the object recognition resources for the deep learning process with a light intensity range of only 23 dB, the prototype machine vision system with the multiple-exposure imaging method demonstrated an object recognition performance with a light intensity range of up to 96 dB.

A Study on Fruit Quality Identification Using YOLO V2 Algorithm

  • Lee, Sang-Hyun
    • International Journal of Advanced Culture Technology
    • /
    • 제9권1호
    • /
    • pp.190-195
    • /
    • 2021
  • Currently, one of the fields leading the 4th industrial revolution is the image recognition field of artificial intelligence, which is showing good results in many fields. In this paper, using is a YOLO V2 model, which is one of the image recognition models, we intend to classify and select into three types according to the characteristics of fruits. To this end, it was designed to proceed the number of iterations of learning 9000 counts based on 640 mandarin image data of 3 classes. For model evaluation, normal, rotten, and unripe mandarin oranges were used based on images. We as a result of the experiment, the accuracy of the learning model was different depending on the number of learning. Normal mandarin oranges showed the highest at 60.5% in 9000 repetition learning, and unripe mandarin oranges also showed the highest at 61.8% in 9000 repetition learning. Lastly, rotten tangerines showed the highest accuracy at 86.0% in 7000 iterations. It will be very helpful if the results of this study are used for fruit farms in rural areas where labor is scarce.

Fire Detection Based on Image Learning by Collaborating CNN-SVM with Enhanced Recall

  • Yongtae Do
    • 센서학회지
    • /
    • 제33권3호
    • /
    • pp.119-124
    • /
    • 2024
  • Effective fire sensing is important to protect lives and property from the disaster. In this paper, we present an intelligent visual sensing method for detecting fires based on machine learning techniques. The proposed method involves a two-step process. In the first step, fire and non-fire images are used to train a convolutional neural network (CNN), and in the next step, feature vectors consisting of 256 values obtained from the CNN are used for the learning of a support vector machine (SVM). Linear and nonlinear SVMs with different parameters are intensively tested. We found that the proposed hybrid method using an SVM with a linear kernel effectively increased the recall rate of fire image detection without compromising detection accuracy when an imbalanced dataset was used for learning. This is a major contribution of this study because recall is important, particularly in the sensing of disaster situations such as fires. In our experiments, the proposed system exhibited an accuracy of 96.9% and a recall rate of 92.9% for test image data.

Use of deep learning in nano image processing through the CNN model

  • Xing, Lumin;Liu, Wenjian;Liu, Xiaoliang;Li, Xin;Wang, Han
    • Advances in nano research
    • /
    • 제12권2호
    • /
    • pp.185-195
    • /
    • 2022
  • Deep learning is another field of artificial intelligence (AI) utilized for computer aided diagnosis (CAD) and image processing in scientific research. Considering numerous mechanical repetitive tasks, reading image slices need time and improper with geographical limits, so the counting of image information is hard due to its strong subjectivity that raise the error ratio in misdiagnosis. Regarding the highest mortality rate of Lung cancer, there is a need for biopsy for determining its class for additional treatment. Deep learning has recently given strong tools in diagnose of lung cancer and making therapeutic regimen. However, identifying the pathological lung cancer's class by CT images in beginning phase because of the absence of powerful AI models and public training data set is difficult. Convolutional Neural Network (CNN) was proposed with its essential function in recognizing the pathological CT images. 472 patients subjected to staging FDG-PET/CT were selected in 2 months prior to surgery or biopsy. CNN was developed and showed the accuracy of 87%, 69%, and 69% in training, validation, and test sets, respectively, for T1-T2 and T3-T4 lung cancer classification. Subsequently, CNN (or deep learning) could improve the CT images' data set, indicating that the application of classifiers is adequate to accomplish better exactness in distinguishing pathological CT images that performs better than few deep learning models, such as ResNet-34, Alex Net, and Dense Net with or without Soft max weights.