• Title/Summary/Keyword: VGG Net

Search Result 96, Processing Time 0.03 seconds

Glaring Wall Pad classification by transfer learning (전이학습을 이용한 전반사가 있는 월패드 분류)

  • Lee, Yong-Jun;Jo, Geun-Sik
    • Proceedings of the Korean Society of Computer Information Conference
    • /
    • 2021.07a
    • /
    • pp.35-36
    • /
    • 2021
  • 딥러닝을 이용한 이미지 처리에서 데이터 셋이 반드시 필요하다. 월패드는 널리 보급되는 다양한 성능을 포함한 IoT가전으로 그 기능의 사용을 돕기 위해서는 해당 월패드에 해당하는 매뉴얼을 제공해야 하고 이를 위해 딥러닝을 이용한 월패드 분류를 이용 할 수 있다. 하지만 월패드 중 일부 모델은 화면의 전반사가 매우 심해 기존의 작은 데이터 셋으로는 딥러닝을 이용한 이미지 분류 성능이 좋지 못하다. 본 논문은 이를 해결하기 위해 추가적으로 데이터 셋을 구축하고 이를 이용해 대규모 데이터로 사전 학습된 VGG16, VGG19, ResNet50, MobileNet 등을 이용해 전이학습을 통해 월패드를 분류한다.

  • PDF

Application and Analysis of Machine Learning for Discriminating Image Copyright (이미지 저작권 판별을 위한 기계학습 적용과 분석)

  • Kim, Sooin;Lee, Sangwoo;Kim, Hakhee;Kim, Wongyum;Hwang, Doosung
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2021.11a
    • /
    • pp.899-902
    • /
    • 2021
  • 본 논문은 이미지 저작권 유무 판별을 분류 문제로 정의하고 기계학습과 합성곱 신경망 모델을 적용하여 해결한다. 학습을 위해 입력 데이터를 고정된 크기로 변환하고 정규화 과정을 수행하여 학습 데이터셋을 준비한다. 저작권 유무 판별 실험에서 SVM, k-NN, 랜덤포레스트, VGG-Net 모델의 분류 성능을 비교 분석한다. VGG-Net C 모델의 결과가 다른 알고리즘과 비교 시 10.65% 높은 성능을 나타냈으며 배치 정규화 층을 이용하여 과적합 현상을 개선했다.

Early Detection of Rice Leaf Blast Disease using Deep-Learning Techniques

  • Syed Rehan Shah;Syed Muhammad Waqas Shah;Hadia Bibi;Mirza Murad Baig
    • International Journal of Computer Science & Network Security
    • /
    • v.24 no.4
    • /
    • pp.211-221
    • /
    • 2024
  • Pakistan is a top producer and exporter of high-quality rice, but traditional methods are still being used for detecting rice diseases. This research project developed an automated rice blast disease diagnosis technique based on deep learning, image processing, and transfer learning with pre-trained models such as Inception V3, VGG16, VGG19, and ResNet50. The modified connection skipping ResNet 50 had the highest accuracy of 99.16%, while the other models achieved 98.16%, 98.47%, and 98.56%, respectively. In addition, CNN and an ensemble model K-nearest neighbor were explored for disease prediction, and the study demonstrated superior performance and disease prediction using recommended web-app approaches.

A Study on the Optimal Convolution Neural Network Backbone for Sinkhole Feature Extraction of GPR B-scan Grayscale Images (GPR B-scan 회색조 이미지의 싱크홀 특성추출 최적 컨볼루션 신경망 백본 연구)

  • Park, Younghoon
    • KSCE Journal of Civil and Environmental Engineering Research
    • /
    • v.44 no.3
    • /
    • pp.385-396
    • /
    • 2024
  • To enhance the accuracy of sinkhole detection using GPR, this study derived a convolutional neural network that can optimally extract sinkhole characteristics from GPR B-scan grayscale images. The pre-trained convolutional neural network is evaluated to be more than twice as effective as the vanilla convolutional neural network. In pre-trained convolutional neural networks, fast feature extraction is found to cause less overfitting than feature extraction. It is analyzed that the top-1 verification accuracy and computation time are different depending on the type of architecture and simulation conditions. Among the pre-trained convolutional neural networks, InceptionV3 are evaluated as most robust for sinkhole detection in GPR B-scan grayscale images. When considering both top-1 verification accuracy and architecture efficiency index, VGG19 and VGG16 are analyzed to have high efficiency as the backbone for extracting sinkhole feature from GPR B-scan grayscale images. MobileNetV3-Large backbone is found to be suitable when mounted on GPR equipment to extract sinkhole feature in real time.

Comparison of Deep Learning Models for Judging Business Card Image Rotation (명함 이미지 회전 판단을 위한 딥러닝 모델 비교)

  • Ji-Hoon, Kyung
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.27 no.1
    • /
    • pp.34-40
    • /
    • 2023
  • A smart business card printing system that automatically prints business cards requested by customers online is being activated. What matters is that the business card submitted by the customer to the system may be abnormal. This paper deals with the problem of determining whether the image of a business card has been abnormally rotated by adopting artificial intelligence technology. It is assumed that the business card rotates 0 degrees, 90 degrees, 180 degrees, and 270 degrees. Experiments were conducted by applying existing VGG, ResNet, and DenseNet artificial neural networks without designing special artificial neural networks, and they were able to distinguish image rotation with an accuracy of about 97%. DenseNet161 showed 97.9% accuracy and ResNet34 also showed 97.2% precision. This illustrates that if the problem is simple, it can produce sufficiently good results even if the neural network is not a complex one.

A study on classification of textile design and extraction of regions of interest (텍스타일 디자인 분류 및 관심 영역 도출에 대한 연구)

  • Chae, Seung Wan;Lee, Woo Chang;Lee, Byoung Woo;Lee, Choong Kwon
    • Smart Media Journal
    • /
    • v.10 no.2
    • /
    • pp.70-75
    • /
    • 2021
  • Grouping and classifying similar designs in design increase efficiency in terms of management and provide convenience in terms of use. Using artificial intelligence algorithms, this study attempted to classify textile designs into four categories: dots, flower patterns, stripes, and geometry. In particular, we explored whether it is possible to find and explain the regions of interest underlying classification from the perspective of artificial intelligence. We randomly extracted a total of 4,536 designs at a ratio of 8:2, comprising 3,629 for training and 907 for testing. The models used in the classification were VGG-16 and ResNet-34, both of which showed excellent classification performance with precision on flower pattern designs of 0.79%, 0.89% and recall of 0.95% and 0.38%. Analysis using the Local Interpretable Model-agnostic Explanation (LIME) technique has shown that geometry and flower-patterned designs derived shapes and petals from the region of interest on which classification was based.

Performance Comparison of the Optimizers in a Faster R-CNN Model for Object Detection of Metaphase Chromosomes (중기 염색체 객체 검출을 위한 Faster R-CNN 모델의 최적화기 성능 비교)

  • Jung, Wonseok;Lee, Byeong-Soo;Seo, Jeongwook
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.23 no.11
    • /
    • pp.1357-1363
    • /
    • 2019
  • In this paper, we compares the performance of the gredient descent optimizers of the Faster Region-based Convolutional Neural Network (R-CNN) model for the chromosome object detection in digital images composed of human metaphase chromosomes. In faster R-CNN, the gradient descent optimizer is used to minimize the objective function of the region proposal network (RPN) module and the classification score and bounding box regression blocks. The gradient descent optimizer. Through performance comparisons among these four gradient descent optimizers in our experiments, we found that the Adamax optimizer could achieve the mean average precision (mAP) of about 52% when considering faster R-CNN with a base network, VGG16. In case of faster R-CNN with a base network, ResNet50, the Adadelta optimizer could achieve the mAP of about 58%.

Performance Comparison of CNN-Based Image Classification Models for Drone Identification System (드론 식별 시스템을 위한 합성곱 신경망 기반 이미지 분류 모델 성능 비교)

  • YeongWan Kim;DaeKyun Cho;GunWoo Park
    • The Journal of the Convergence on Culture Technology
    • /
    • v.10 no.4
    • /
    • pp.639-644
    • /
    • 2024
  • Recent developments in the use of drones on battlefields, extending beyond reconnaissance to firepower support, have greatly increased the importance of technologies for early automatic drone identification. In this study, to identify an effective image classification model that can distinguish drones from other aerial targets of similar size and appearance, such as birds and balloons, we utilized a dataset of 3,600 images collected from the internet. We adopted a transfer learning approach that combines the feature extraction capabilities of three pre-trained convolutional neural network models (VGG16, ResNet50, InceptionV3) with an additional classifier. Specifically, we conducted a comparative analysis of the performance of these three pre-trained models to determine the most effective one. The results showed that the InceptionV3 model achieved the highest accuracy at 99.66%. This research represents a new endeavor in utilizing existing convolutional neural network models and transfer learning for drone identification, which is expected to make a significant contribution to the advancement of drone identification technologies.

A Defect Detection Algorithm of Denim Fabric Based on Cascading Feature Extraction Architecture

  • Shuangbao, Ma;Renchao, Zhang;Yujie, Dong;Yuhui, Feng;Guoqin, Zhang
    • Journal of Information Processing Systems
    • /
    • v.19 no.1
    • /
    • pp.109-117
    • /
    • 2023
  • Defect detection is one of the key factors in fabric quality control. To improve the speed and accuracy of denim fabric defect detection, this paper proposes a defect detection algorithm based on cascading feature extraction architecture. Firstly, this paper extracts these weight parameters of the pre-trained VGG16 model on the large dataset ImageNet and uses its portability to train the defect detection classifier and the defect recognition classifier respectively. Secondly, retraining and adjusting partial weight parameters of the convolution layer were retrained and adjusted from of these two training models on the high-definition fabric defect dataset. The last step is merging these two models to get the defect detection algorithm based on cascading architecture. Then there are two comparative experiments between this improved defect detection algorithm and other feature extraction methods, such as VGG16, ResNet-50, and Xception. The results of experiments show that the defect detection accuracy of this defect detection algorithm can reach 94.3% and the speed is also increased by 1-3 percentage points.

Construction of a Bark Dataset for Automatic Tree Identification and Developing a Convolutional Neural Network-based Tree Species Identification Model (수목 동정을 위한 수피 분류 데이터셋 구축과 합성곱 신경망 기반 53개 수종의 동정 모델 개발)

  • Kim, Tae Kyung;Baek, Gyu Heon;Kim, Hyun Seok
    • Journal of Korean Society of Forest Science
    • /
    • v.110 no.2
    • /
    • pp.155-164
    • /
    • 2021
  • Many studies have been conducted on developing automatic plant identification algorithms using machine learning to various plant features, such as leaves and flowers. Unlike other plant characteristics, barks show only little change regardless of the season and are maintained for a long period. Nevertheless, barks show a complex shape with a large variation depending on the environment, and there are insufficient materials that can be utilized to train algorithms. Here, in addition to the previously published bark image dataset, BarkNet v.1.0, images of barks were collected, and a dataset consisting of 53 tree species that can be easily observed in Korea was presented. A convolutional neural network (CNN) was trained and tested on the dataset, and the factors that interfere with the model's performance were identified. For CNN architecture, VGG-16 and 19 were utilized. As a result, VGG-16 achieved 90.41% and VGG-19 achieved 92.62% accuracy. When tested on new tree images that do not exist in the original dataset but belong to the same genus or family, it was confirmed that more than 80% of cases were successfully identified as the same genus or family. Meanwhile, it was found that the model tended to misclassify when there were distracting features in the image, including leaves, mosses, and knots. In these cases, we propose that random cropping and classification by majority votes are valid for improving possible errors in training and inferences.