• Title/Summary/Keyword: ImageNet

Search Result 806, Processing Time 0.022 seconds

Design of Conceptual Image Annotation System Using WordNet (WordNet 기반 개념적 이미지 주석 시스템 설계)

  • 조미영;최준호;김판구
    • Proceedings of the Korea Multimedia Society Conference
    • /
    • 2002.05d
    • /
    • pp.1081-1086
    • /
    • 2002
  • 이미지검색을 위해서 객체의 시각적인 특징에 대한 저차원의 특징 정보를 추출하고 이미지에 의미를 부여하기 위하여 주석을 다는 것이 일반적이다. 하지만 주석 기반 검색에서는 주석으로 달아 놓은 단어와 정확한 매칭이 없다면 찾을 수가 없다. 이러한 문제를 해결하기 위해 재질의 질의어 확장과 같은 기법을 써서 문제를 해결해 왔으나 여전히 개념적 매칭이 아닌 스트링 매칭의 문제를 안고 있다고 볼 수 있다. 이에 본 논문에서는 이미지 관련 Text에서 단어를 추출한 후 추출된 단어들간의 개념 관계를 WordNet을 이용하여 표현한 주석 시스템을 제안한다. 이 시스템은 단순 스트링 매칭이 아닌 개념적 매칭에 의한 개념 기반 검색을 지원할 수 있다.

  • PDF

Improvement of Active Net model for Region Detection in an Image (개선된 Active Net Model을 이용한 이미지 영역검출)

  • 남기환;배철수;설증보;나상동
    • Proceedings of the Korea Multimedia Society Conference
    • /
    • 2004.05a
    • /
    • pp.243-246
    • /
    • 2004
  • 본 논문은 영상인식 방법으로 개선된 Active Model을 이용한 방법을 제안한다. 제안된 방법은 모든 격자 블록 영역이 동일한 구조를 가지며, 기존의 Active net에서 문제가 되었던 목표물을 탐지하는 능력이 개선되었다. 실험 결과로서 제안된 방법이 수직, 수평 방향에서 목표물 포착에 효과적임을 보여주었으며, 실제 도로 영상에 적용한 결과 제안한 방법의 효율성을 입증할 수 있었다.

  • PDF

Faster D2-Net for Screen Image Matching (스크린 이미지 매칭을 위한 Faster D2-Net)

  • Chun, Hye-Won;Han, Seong-Soo;Jeong, Chang-Sung
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2021.05a
    • /
    • pp.429-432
    • /
    • 2021
  • 스마트 기기와 애플리케이션의 테스트를 위해 빠르고 정확하게 스마트 기기 화면 상에서 테스트가 필요한 위치를 추출해야 한다. 필요한 위치를 추출할 때 스마트 기기 화면과 테스트할 수 있는 영역의 매칭 방식을 사용하는데 이를 위해 이미지의 변형이 발생해도 원하는 영역의 matching point 을 빠르고 정확하게 추출하는 feature matching 방식의 D2-Net 의 feature extraction 모델과 fitting algorithm 을 변경하였다.

Mushroom Image Recognition using Convolutional Neural Network and Transfer Learning (컨볼루션 신경망과 전이 학습을 이용한 버섯 영상 인식)

  • Kang, Euncheol;Han, Yeongtae;Oh, Il-Seok
    • KIISE Transactions on Computing Practices
    • /
    • v.24 no.1
    • /
    • pp.53-57
    • /
    • 2018
  • A poisoning accident is often caused by a situation in which people eat poisonous mushrooms because they cannot distinguish between edible mushrooms and poisonous mushrooms. In this paper, we propose an automatic mushroom recognition system by using the convolutional neural network. We collected 1478 mushroom images of 38 species using image crawling, and used the dataset for learning the convolutional neural network. A comparison experiment using AlexNet, VGGNet, and GoogLeNet was performed using the collected datasets, and a comparison experiment using a class number expansion and a fine-tuning technique for transfer learning were performed. As a result of our experiment, we achieve 82.63% top-1 accuracy and 96.84% top-5 accuracy on test set of our dataset.

Study on the Surface Defect Classification of Al 6061 Extruded Material By Using CNN-Based Algorithms (CNN을 이용한 Al 6061 압출재의 표면 결함 분류 연구)

  • Kim, S.B.;Lee, K.A.
    • Transactions of Materials Processing
    • /
    • v.31 no.4
    • /
    • pp.229-239
    • /
    • 2022
  • Convolution Neural Network(CNN) is a class of deep learning algorithms and can be used for image analysis. In particular, it has excellent performance in finding the pattern of images. Therefore, CNN is commonly applied for recognizing, learning and classifying images. In this study, the surface defect classification performance of Al 6061 extruded material using CNN-based algorithms were compared and evaluated. First, the data collection criteria were suggested and a total of 2,024 datasets were prepared. And they were randomly classified into 1,417 learning data and 607 evaluation data. After that, the size and quality of the training data set were improved using data augmentation techniques to increase the performance of deep learning. The CNN-based algorithms used in this study were VGGNet-16, VGGNet-19, ResNet-50 and DenseNet-121. The evaluation of the defect classification performance was made by comparing the accuracy, loss, and learning speed using verification data. The DenseNet-121 algorithm showed better performance than other algorithms with an accuracy of 99.13% and a loss value of 0.037. This was due to the structural characteristics of the DenseNet model, and the information loss was reduced by acquiring information from all previous layers for image identification in this algorithm. Based on the above results, the possibility of machine vision application of CNN-based model for the surface defect classification of Al extruded materials was also discussed.

Tomato Crop Diseases Classification Models Using Deep CNN-based Architectures (심층 CNN 기반 구조를 이용한 토마토 작물 병해충 분류 모델)

  • Kim, Sam-Keun;Ahn, Jae-Geun
    • Journal of the Korea Academia-Industrial cooperation Society
    • /
    • v.22 no.5
    • /
    • pp.7-14
    • /
    • 2021
  • Tomato crops are highly affected by tomato diseases, and if not prevented, a disease can cause severe losses for the agricultural economy. Therefore, there is a need for a system that quickly and accurately diagnoses various tomato diseases. In this paper, we propose a system that classifies nine diseases as well as healthy tomato plants by applying various pretrained deep learning-based CNN models trained on an ImageNet dataset. The tomato leaf image dataset obtained from PlantVillage is provided as input to ResNet, Xception, and DenseNet, which have deep learning-based CNN architectures. The proposed models were constructed by adding a top-level classifier to the basic CNN model, and they were trained by applying a 5-fold cross-validation strategy. All three of the proposed models were trained in two stages: transfer learning (which freezes the layers of the basic CNN model and then trains only the top-level classifiers), and fine-tuned learning (which sets the learning rate to a very small number and trains after unfreezing basic CNN layers). SGD, RMSprop, and Adam were applied as optimization algorithms. The experimental results show that the DenseNet CNN model to which the RMSprop algorithm was applied output the best results, with 98.63% accuracy.

Substitutability of Noise Reduction Algorithm based Conventional Thresholding Technique to U-Net Model for Pancreas Segmentation (이자 분할을 위한 노이즈 제거 알고리즘 기반 기존 임계값 기법 대비 U-Net 모델의 대체 가능성)

  • Sewon Lim;Youngjin Lee
    • Journal of the Korean Society of Radiology
    • /
    • v.17 no.5
    • /
    • pp.663-670
    • /
    • 2023
  • In this study, we aimed to perform a comparative evaluation using quantitative factors between a region-growing based segmentation with noise reduction algorithms and a U-Net based segmentation. Initially, we applied median filter, median modified Wiener filter, and fast non-local means algorithm to computed tomography (CT) images, followed by region-growing based segmentation. Additionally, we trained a U-Net based segmentation model to perform segmentation. Subsequently, to compare and evaluate the segmentation performance of cases with noise reduction algorithms and cases with U-Net, we measured root mean square error (RMSE) and peak signal to noise ratio (PSNR), universal quality image index (UQI), and dice similarity coefficient (DSC). The results showed that using U-Net for segmentation yielded the most improved performance. The values of RMSE, PSNR, UQI, and DSC were measured as 0.063, 72.11, 0.841, and 0.982 respectively, which indicated improvements of 1.97, 1.09, 5.30, and 1.99 times compared to noisy images. In conclusion, U-Net proved to be effective in enhancing segmentation performance compared to noise reduction algorithms in CT images.

Keyword Selection for Visual Search based on Wikipedia (비주얼 검색을 위한 위키피디아 기반의 질의어 추출)

  • Kim, Jongwoo;Cho, Soosun
    • Journal of Korea Multimedia Society
    • /
    • v.21 no.8
    • /
    • pp.960-968
    • /
    • 2018
  • The mobile visual search service uses a query image to acquire linkage information through pre-constructed DB search. From the standpoint of this purpose, it would be more useful if you could perform a search on a web-based keyword search system instead of a pre-built DB search. In this paper, we propose a representative query extraction algorithm to be used as a keyword on a web-based search system. To do this, we use image classification labels generated by the CNN (Convolutional Neural Network) algorithm based on Deep Learning, which has a remarkable performance in image recognition. In the query extraction algorithm, dictionary meaningful words are extracted using Wikipedia, and hierarchical categories are constructed using WordNet. The performance of the proposed algorithm is evaluated by measuring the system response time.

High-performance of Deep learning Colorization With Wavelet fusion (웨이블릿 퓨전에 의한 딥러닝 색상화의 성능 향상)

  • Kim, Young-Back;Choi, Hyun;Cho, Joong-Hwee
    • IEMEK Journal of Embedded Systems and Applications
    • /
    • v.13 no.6
    • /
    • pp.313-319
    • /
    • 2018
  • We propose a post-processing algorithm to improve the quality of the RGB image generated by deep learning based colorization from the gray-scale image of an infrared camera. Wavelet fusion is used to generate a new luminance component of the RGB image luminance component from the deep learning model and the luminance component of the infrared camera. PSNR is increased for all experimental images by applying the proposed algorithm to RGB images generated by two deep learning models of SegNet and DCGAN. For the SegNet model, the average PSNR is improved by 1.3906dB at level 1 of the Haar wavelet method. For the DCGAN model, PSNR is improved 0.0759dB on the average at level 5 of the Daubechies wavelet method. It is also confirmed that the edge components are emphasized by the post-processing and the visibility is improved.

Automatic Volumetric Brain Tumor Segmentation using Convolutional Neural Networks

  • Yavorskyi, Vladyslav;Sull, Sanghoon
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2019.05a
    • /
    • pp.432-435
    • /
    • 2019
  • Convolutional Neural Networks (CNNs) have recently been gaining popularity in the medical image analysis field because of their image segmentation capabilities. In this paper, we present a CNN that performs automated brain tumor segmentations of sparsely annotated 3D Magnetic Resonance Imaging (MRI) scans. Our CNN is based on 3D U-net architecture, and it includes separate Dilated and Depth-wise Convolutions. It is fully-trained on the BraTS 2018 data set, and it produces more accurate results even when compared to the winners of the BraTS 2017 competition despite having a significantly smaller amount of parameters.

  • PDF