• Title/Summary/Keyword: inception-V3

Search Result 77, Processing Time 0.032 seconds

3D Res-Inception Network Transfer Learning for Multiple Label Crowd Behavior Recognition

  • Nan, Hao;Li, Min;Fan, Lvyuan;Tong, Minglei
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.13 no.3
    • /
    • pp.1450-1463
    • /
    • 2019
  • The problem towards crowd behavior recognition in a serious clustered scene is extremely challenged on account of variable scales with non-uniformity. This paper aims to propose a crowed behavior classification framework based on a transferring hybrid network blending 3D res-net with inception-v3. First, the 3D res-inception network is presented so as to learn the augmented visual feature of UCF 101. Then the target dataset is applied to fine-tune the network parameters in an attempt to classify the behavior of densely crowded scenes. Finally, a transferred entropy function is used to calculate the probability of multiple labels in accordance with these features. Experimental results show that the proposed method could greatly improve the accuracy of crowd behavior recognition and enhance the accuracy of multiple label classification.

Breast Cancer Histopathological Image Classification Based on Deep Neural Network with Pre-Trained Model Architecture (사전훈련된 모델구조를 이용한 심층신경망 기반 유방암 조직병리학적 이미지 분류)

  • Mudeng, Vicky;Lee, Eonjin;Choe, Se-woon
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2022.05a
    • /
    • pp.399-401
    • /
    • 2022
  • A definitive diagnosis to classify the breast malignancy status may be achieved by microscopic analysis using surgical open biopsy. However, this procedure requires experts in the specializing of histopathological image analysis directing to time-consuming and high cost. To overcome these issues, deep learning is considered practically efficient to categorize breast cancer into benign and malignant from histopathological images in order to assist pathologists. This study presents a pre-trained convolutional neural network model architecture with a 100% fine-tuning scheme and Adagrad optimizer to classify the breast cancer histopathological images into benign and malignant using a 40× magnification BreaKHis dataset. The pre-trained architecture was constructed using the InceptionResNetV2 model to generate a modified InceptionResNetV2 by substituting the last layer with dense and dropout layers. The results by demonstrating training loss of 0.25%, training accuracy of 99.96%, validation loss of 3.10%, validation accuracy of 99.41%, test loss of 8.46%, and test accuracy of 98.75% indicated that the modified InceptionResNetV2 model is reliable to predict the breast malignancy type from histopathological images. Future works are necessary to focus on k-fold cross-validation, optimizer, model, hyperparameter optimization, and classification on 100×, 200×, and 400× magnification.

  • PDF

Classification of Raccoon dog and Raccoon with Transfer Learning and Data Augmentation (전이 학습과 데이터 증강을 이용한 너구리와 라쿤 분류)

  • Dong-Min Park;Yeong-Seok Jo;Seokwon Yeom
    • Journal of the Institute of Convergence Signal Processing
    • /
    • v.24 no.1
    • /
    • pp.34-41
    • /
    • 2023
  • In recent years, as the range of human activities has increased, the introduction of alien species has become frequent. Among them, raccoons have been designated as harmful animals since 2020. Raccoons are similar in size and shape to raccoon dogs, so they generally need to be distinguished in capturing them. To solve this problem, we use VGG19, ResNet152V2, InceptionV3, InceptionResNet and NASNet, which are CNN deep learning models specialized for image classification. The parameters to be used for learning are pre-trained with a large amount of data, ImageNet. In order to classify the raccoon and raccoon dog datasets as outward features of animals, the image was converted to grayscale and brightness was normalized. Augmentation methods were applied using left and right inversion, rotation, scaling, and shift to create sufficient data for transfer learning. The FCL consists of 1 layer for the non-augmented dataset while 4 layers for the augmented dataset. Comparing the accuracy of various augmented datasets, the performance increased as more augmentation methods were applied.

Dog-Species Classification through CycleGAN and Standard Data Augmentation

  • Chan, Park;Nammee, Moon
    • Journal of Information Processing Systems
    • /
    • v.19 no.1
    • /
    • pp.67-79
    • /
    • 2023
  • In the image field, data augmentation refers to increasing the amount of data through an editing method such as rotating or cropping a photo. In this study, a generative adversarial network (GAN) image was created using CycleGAN, and various colors of dogs were reflected through data augmentation. In particular, dog data from the Stanford Dogs Dataset and Oxford-IIIT Pet Dataset were used, and 10 breeds of dog, corresponding to 300 images each, were selected. Subsequently, a GAN image was generated using CycleGAN, and four learning groups were established: 2,000 original photos (group I); 2,000 original photos + 1,000 GAN images (group II); 3,000 original photos (group III); and 3,000 original photos + 1,000 GAN images (group IV). The amount of data in each learning group was augmented using existing data augmentation methods such as rotating, cropping, erasing, and distorting. The augmented photo data were used to train the MobileNet_v3_Large, ResNet-152, InceptionResNet_v2, and NASNet_Large frameworks to evaluate the classification accuracy and loss. The top-3 accuracy for each deep neural network model was as follows: MobileNet_v3_Large of 86.4% (group I), 85.4% (group II), 90.4% (group III), and 89.2% (group IV); ResNet-152 of 82.4% (group I), 83.7% (group II), 84.7% (group III), and 84.9% (group IV); InceptionResNet_v2 of 90.7% (group I), 88.4% (group II), 93.3% (group III), and 93.1% (group IV); and NASNet_Large of 85% (group I), 88.1% (group II), 91.8% (group III), and 92% (group IV). The InceptionResNet_v2 model exhibited the highest image classification accuracy, and the NASNet_Large model exhibited the highest increase in the accuracy owing to data augmentation.

Study of the Cavitation Inception Speed (CIS) Improvement Through the Propeller Design and the Stern Appendage Modification (프로펠러 설계 및 선미 부가물 수정에 따른 캐비테이션 초기발생 선속(CIS) 성능 향상 연구)

  • Jong-Woo Ahn;Gun-Do Kim;Bu-Geun Paik;Young-Ha Park;Han-Shin Seol
    • Journal of the Society of Naval Architects of Korea
    • /
    • v.60 no.4
    • /
    • pp.231-239
    • /
    • 2023
  • In order to improve the propeller Cavitation Inception Speed (CIS) performance, it needs to modify the propeller geometry and the wake distribution that flows into the propeller. In the previous study, the twisted angles of the V-strut were modified to improve propeller CIS, cavitation behavior and pressure fluctuation performances. Then the propeller behind the modified V-strut (New strut) showed better cavitation characteristics than that behind the existing V-strut (Old strut). However, the CIS of Suction Side Tip Vortex (SSTV) and Pressure Side Tp Vortex (PSTV) showed a big difference at behind each V-strut. In this study, the balance design is conducted to minimize the difference between SSTV CIS and PSTV CIS at behind each V-strut. To improve the propeller CIS performance, 1 propeller is designed at behind the old strut and 3 propellers are designed at behind the new strut. The propeller CIS is increased through the balance design and the stern appendage modification. The final propeller CIS is increased about 5.3 knots higher than that of the existing propeller at behind the old strut. On the basis of the present study, it is thought that the better improvement method for the propeller CIS would be suggested.

Enhancement of Tongue Segmentation by Using Data Augmentation (데이터 증강을 이용한 혀 영역 분할 성능 개선)

  • Chen, Hong;Jung, Sung-Tae
    • The Journal of Korea Institute of Information, Electronics, and Communication Technology
    • /
    • v.13 no.5
    • /
    • pp.313-322
    • /
    • 2020
  • A large volume of data will improve the robustness of deep learning models and avoid overfitting problems. In automatic tongue segmentation, the availability of annotated tongue images is often limited because of the difficulty of collecting and labeling the tongue image datasets in reality. Data augmentation can expand the training dataset and increase the diversity of training data by using label-preserving transformations without collecting new data. In this paper, augmented tongue image datasets were developed using seven augmentation techniques such as image cropping, rotation, flipping, color transformations. Performance of the data augmentation techniques were studied using state-of-the-art transfer learning models, for instance, InceptionV3, EfficientNet, ResNet, DenseNet and etc. Our results show that geometric transformations can lead to more performance gains than color transformations and the segmentation accuracy can be increased by 5% to 20% compared with no augmentation. Furthermore, a random linear combination of geometric and color transformations augmentation dataset gives the superior segmentation performance than all other datasets and results in a better accuracy of 94.98% with InceptionV3 models.

Deep Learning Based Tree Recognition rate improving Method for Elementary and Middle School Learning

  • Choi, Jung-Eun;Yong, Hwan-Seung
    • Journal of the Korea Society of Computer and Information
    • /
    • v.24 no.12
    • /
    • pp.9-16
    • /
    • 2019
  • The goal of this study is to propose an efficient model for recognizing and classifying tree images to measure the accuracy that can be applied to smart devices during class. From the 2009 revised textbook to the 2015 revised textbook, the learning objective to the fourth-grade science textbook of elementary schools was added to the plant recognition utilizing smart devices. In this study, we compared the recognition rates of trees before and after retraining using a pre-trained inception V3 model, which is the support of the Google Inception V3. In terms of tree recognition, it can distinguish several features, including shapes, bark, leaves, flowers, and fruits that may lead to the recognition rate. Furthermore, if all the leaves of trees may fall during winter, it may challenge to identify the type of tree, as only the bark of the tree will remain some leaves. Therefore, the effective tree classification model is presented through the combination of the images by tree type and the method of combining the model for the accuracy of each tree type. I hope that this model will apply to smart devices used in educational settings.

Implementation of Seed Germination Confirmation System with Deep Learning (딥 러닝을 활용한 씨앗 발아 확인 시스템)

  • Gim, U Ju;Kwon, Min Seo;Lee, Jae Jun;Yoo, Kwan Hee;Hong, Jang-Eui;Nasridinov, Aziz
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2018.10a
    • /
    • pp.603-605
    • /
    • 2018
  • 최근 대두되고 있는 딥 러닝은 학습을 통해 사물이나 데이터를 군집화하거나 분류하는 데 사용하는 기술이다. 본 논문은 딥 러닝에 활용하기 위해 개발된 오픈소스 소프트웨어인 텐서플로 Inception V3을 사용해 연구를 진행했다. 딥 러닝을 활용한 씨앗 발아 확인 시스템은 기존의 영상 처리를 활용한 시스템에서 고안했으며, 씨앗 발아 여부의 정확성이 떨어지는 단점을 개선하고, 모든 종자들의 발아 여부를 확인할 수 있도록 구현해 사용자가 효과적으로 연구를 수행할 수 있도록 하는 목적에 있다.

Comparison of Image Classification Performance in Convolutional Neural Network according to Transfer Learning (전이학습에 방법에 따른 컨벌루션 신경망의 영상 분류 성능 비교)

  • Park, Sung-Wook;Kim, Do-Yeon
    • Journal of Korea Multimedia Society
    • /
    • v.21 no.12
    • /
    • pp.1387-1395
    • /
    • 2018
  • Core algorithm of deep learning Convolutional Neural Network(CNN) shows better performance than other machine learning algorithms. However, if there is not sufficient data, CNN can not achieve satisfactory performance even if the classifier is excellent. In this situation, it has been proven that the use of transfer learning can have a great effect. In this paper, we apply two transition learning methods(freezing, retraining) to three CNN models(ResNet-50, Inception-V3, DenseNet-121) and compare and analyze how the classification performance of CNN changes according to the methods. As a result of statistical significance test using various evaluation indicators, ResNet-50, Inception-V3, and DenseNet-121 differed by 1.18 times, 1.09 times, and 1.17 times, respectively. Based on this, we concluded that the retraining method may be more effective than the freezing method in case of transition learning in image classification problem.

Recommended System for Cosmetics Using Inception v3 module (Inception v3를 이용한 화장품 추천 시스템)

  • Jang, YoungHoon;Raza, Syed Muhammad;Kim, MoonSeong;Choo, HyunSeung
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2020.05a
    • /
    • pp.372-374
    • /
    • 2020
  • 최근 화장품이나 뷰티산업의 성장이 가속화되고 있다. 이에 따라 시장에 다양한 뷰티제품들이 출시되고 있지만 그로 인해 오히려 본인에게 적합한 제품이 무엇인지 알지 못하는 경우가 많다. 온라인을 통해 구매하는 경우 구매후기 및 광고에 의지해야 하며 전문가의 조언을 구하기 위해서는 오프라인 상점을 방문할 수밖에 없다. 그러나 오프라인 상점을 방문한 경우에도 자신에게 적합한 화장품을 추천받는 것 또한 다분하지 않다. 본 논문에서는 이러한 문제점을 해결하고자 온라인 환경에서 소비자에게 맞는 상품의 광고 및 정보를 받을 수 있는 화장품 추천 서비스를 제안한다. 또한 제안서비스는 AI기능을 적용하여 기존의 방식보다 소비자 친화적인 서바스를 제공하는 것을 목표로 한다.