• Title/Summary/Keyword: CNN Model

Search Result 963, Processing Time 0.029 seconds

Discriminant analysis of grain flours for rice paper using fluorescence hyperspectral imaging system and chemometric methods

  • Seo, Youngwook;Lee, Ahyeong;Kim, Bal-Geum;Lim, Jongguk
    • Korean Journal of Agricultural Science
    • /
    • v.47 no.3
    • /
    • pp.633-644
    • /
    • 2020
  • Rice paper is an element of Vietnamese cuisine that can be used to wrap vegetables and meat. Rice and starch are the main ingredients of rice paper and their mixing ratio is important for quality control. In a commercial factory, assessment of food safety and quantitative supply is a challenging issue. A rapid and non-destructive monitoring system is therefore necessary in commercial production systems to ensure the food safety of rice and starch flour for the rice paper wrap. In this study, fluorescence hyperspectral imaging technology was applied to classify grain flours. Using the 3D hyper cube of fluorescence hyperspectral imaging (fHSI, 420 - 730 nm), spectral and spatial data and chemometric methods were applied to detect and classify flours. Eight flours (rice: 4, starch: 4) were prepared and hyperspectral images were acquired in a 5 (L) × 5 (W) × 1.5 (H) cm container. Linear discriminant analysis (LDA), partial least square discriminant analysis (PLSDA), support vector machine (SVM), classification and regression tree (CART), and random forest (RF) with a few preprocessing methods (multivariate scatter correction [MSC], 1st and 2nd derivative and moving average) were applied to classify grain flours and the accuracy was compared using a confusion matrix (accuracy and kappa coefficient). LDA with moving average showed the highest accuracy at A = 0.9362 (K = 0.9270). 1D convolutional neural network (CNN) demonstrated a classification result of A = 0.94 and showed improved classification results between mimyeon flour (MF)1 and MF2 of 0.72 and 0.87, respectively. In this study, the potential of non-destructive detection and classification of grain flours using fHSI technology and machine learning methods was demonstrated.

Low Resolution Infrared Image Deep Convolution Neural Network for Embedded System

  • Hong, Yong-hee;Jin, Sang-hun;Kim, Dae-hyeon;Jhee, Ho-Jin
    • Journal of the Korea Society of Computer and Information
    • /
    • v.26 no.6
    • /
    • pp.1-8
    • /
    • 2021
  • In this paper, we propose reinforced VGG style network structure for low performance embedded system to classify low resolution infrared image. The combination of reinforced VGG style network structure and global average pooling makes lower computational complexity and higher accuracy. The proposed method classify the synthesize image which have 9 class 3,723,328ea images made from OKTAL-SE tool. The reinforced VGG style network structure composed of 4 filters on input and 16 filters on output from max pooling layer shows about 34% lower computational complexity and about 2.4% higher accuracy then the first parameter minimized network structure made for embedded system composed of 8 filters on input and 8 filters on output from max pooling layer. Finally we get 96.1% accuracy model. Additionally we confirmed the about 31% lower inference lead time in ported C code.

Using the Deep Learning for the System Architecture of Image Prediction (엔터프라이즈 환경의 딥 러닝을 활용한 이미지 예측 시스템 아키텍처)

  • Cheon, Eun Young;Choi, Sung-Ja
    • Journal of Digital Convergence
    • /
    • v.17 no.10
    • /
    • pp.259-264
    • /
    • 2019
  • This paper proposes an image prediction system architecture for deep running in enterprise environment. Easily transform into an artificial intelligence platform for an enterprise environment, and allow sufficient deep-running services to be developed and modified even in Java-centric architectures to improve the shortcomings of Java-centric enterprise development because artificial intelligence platforms are concentrated in the pipeline. In addition, based on the proposed environment, we propose a more accurate prediction system in the deep running architecture environment that has been previously learned through image forecasting experiments. Experiments show 95.23% accuracy in the image example provided for deep running to be performed, and the proposed model shows 96.54% accuracy compared to other similar models.

Courses Recommendation Algorithm Based On Performance Prediction In E-Learning

  • Koffi, Dagou Dangui Augustin Sylvain Legrand;Ouattara, Nouho;Mambe, Digrais Moise;Oumtanaga, Souleymane;ADJE, Assohoun
    • International Journal of Computer Science & Network Security
    • /
    • v.21 no.2
    • /
    • pp.148-157
    • /
    • 2021
  • The effectiveness of recommendation systems depends on the performance of the algorithms with which these systems are designed. The quality of the algorithms themselves depends on the quality of the strategies with which they were designed. These strategies differ from author to author. Thus, designing a good recommendation system means implementing the good strategies. It's in this context that several research works have been proposed on various strategies applied to algorithms to meet the needs of recommendations. Researchers are trying indefinitely to address this objective of seeking the qualities of recommendation algorithms. In this paper, we propose a new algorithm for recommending learning items. Learner performance predictions and collaborative recommendation methods are used as strategies for this algorithm. The proposed performance prediction model is based on convolutional neural networks (CNN). The results of the performance predictions are used by the proposed recommendation algorithm. The results of the predictions obtained show the efficiency of Deep Learning compared to the k-nearest neighbor (k-NN) algorithm. The proposed recommendation algorithm improves the recommendations of the learners' learning items. This algorithm also has the particularity of dissuading learning items in the learner's profile that are deemed inadequate for his or her training.

The Effect of Type of Input Image on Accuracy in Classification Using Convolutional Neural Network Model (컨볼루션 신경망 모델을 이용한 분류에서 입력 영상의 종류가 정확도에 미치는 영향)

  • Kim, Min Jeong;Kim, Jung Hun;Park, Ji Eun;Jeong, Woo Yeon;Lee, Jong Min
    • Journal of Biomedical Engineering Research
    • /
    • v.42 no.4
    • /
    • pp.167-174
    • /
    • 2021
  • The purpose of this study is to classify TIFF images, PNG images, and JPEG images using deep learning, and to compare the accuracy by verifying the classification performance. The TIFF, PNG, and JPEG images converted from chest X-ray DICOM images were applied to five deep neural network models performed in image recognition and classification to compare classification performance. The data consisted of a total of 4,000 X-ray images, which were converted from DICOM images into 16-bit TIFF images and 8-bit PNG and JPEG images. The learning models are CNN models - VGG16, ResNet50, InceptionV3, DenseNet121, and EfficientNetB0. The accuracy of the five convolutional neural network models of TIFF images is 99.86%, 99.86%, 99.99%, 100%, and 99.89%. The accuracy of PNG images is 99.88%, 100%, 99.97%, 99.87%, and 100%. The accuracy of JPEG images is 100%, 100%, 99.96%, 99.89%, and 100%. Validation of classification performance using test data showed 100% in accuracy, precision, recall and F1 score. Our classification results show that when DICOM images are converted to TIFF, PNG, and JPEG images and learned through preprocessing, the learning works well in all formats. In medical imaging research using deep learning, the classification performance is not affected by converting DICOM images into any format.

Damage localization and quantification of a truss bridge using PCA and convolutional neural network

  • Jiajia, Hao;Xinqun, Zhu;Yang, Yu;Chunwei, Zhang;Jianchun, Li
    • Smart Structures and Systems
    • /
    • v.30 no.6
    • /
    • pp.673-686
    • /
    • 2022
  • Deep learning algorithms for Structural Health Monitoring (SHM) have been extracting the interest of researchers and engineers. These algorithms commonly used loss functions and evaluation indices like the mean square error (MSE) which were not originally designed for SHM problems. An updated loss function which was specifically constructed for deep-learning-based structural damage detection problems has been proposed in this study. By tuning the coefficients of the loss function, the weights for damage localization and quantification can be adapted to the real situation and the deep learning network can avoid unnecessary iterations on damage localization and focus on the damage severity identification. To prove efficiency of the proposed method, structural damage detection using convolutional neural networks (CNNs) was conducted on a truss bridge model. Results showed that the validation curve with the updated loss function converged faster than the traditional MSE. Data augmentation was conducted to improve the anti-noise ability of the proposed method. For reducing the training time, the normalized modal strain energy change (NMSEC) was extracted, and the principal component analysis (PCA) was adopted for dimension reduction. The results showed that the training time was reduced by 90% and the damage identification accuracy could also have a slight increase. Furthermore, the effect of different modes and elements on the training dataset was also analyzed. The proposed method could greatly improve the performance for structural damage detection on both the training time and detection accuracy.

Development of a Deep Learning Algorithm for Anomaly Detection of Manufacturing Facility (설비 이상탐지를 위한 딥러닝 알고리즘 개발)

  • Kim, Min-Hee;Jin, Kyo-Hong
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.26 no.2
    • /
    • pp.199-206
    • /
    • 2022
  • A malfunction or breakdown of a manufacturing facility leads to product defects and the suspension of production lines, resulting in huge financial losses for manufacturers. Due to the spread of smart factory services, a large amount of data is being collected in factories, and AI-based research is being conducted to predict and diagnose manufacturing facility breakdowns or manufacturing site efficiency. However, because of the characteristics of manufacturing data, such as a severe class imbalance about abnormalities and ambiguous label information that distinguishes abnormalities, developing classification or anomaly detection models is highly difficult. In this paper, we present an deep learning algorithm for anomaly detection of a manufacturing facility using reconstruction loss of CNN-based model and ananlyze its performance. The algorithm detects anomalies by relying solely on normal data from the facility's manufacturing data in the exclusion of abnormal data.

Implementation of an alarm system with AI image processing to detect whether a helmet is worn or not and a fall accident (헬멧 착용 여부 및 쓰러짐 사고 감지를 위한 AI 영상처리와 알람 시스템의 구현)

  • Yong-Hwa Jo;Hyuek-Jae Lee
    • Journal of the Institute of Convergence Signal Processing
    • /
    • v.23 no.3
    • /
    • pp.150-159
    • /
    • 2022
  • This paper presents an implementation of detecting whether a helmet is worn and there is a fall accident through individual image analysis in real-time from extracting the image objects of several workers active in the industrial field. In order to detect image objects of workers, YOLO, a deep learning-based computer vision model, was used, and for whether a helmet is worn or not, the extracted images with 5,000 different helmet learning data images were applied. For whether a fall accident occurred, the position of the head was checked using the Pose real-time body tracking algorithm of Mediapipe, and the movement speed was calculated to determine whether the person fell. In addition, to give reliability to the result of a falling accident, a method to infer the posture of an object by obtaining the size of YOLO's bounding box was proposed and implemented. Finally, Telegram API Bot and Firebase DB server were implemented for notification service to administrators.

An EEG-fNIRS Hybridization Technique in the Multi-class Classification of Alzheimer's Disease Facilitated by Machine Learning (기계학습 기반 알츠하이머성 치매의 다중 분류에서 EEG-fNIRS 혼성화 기법)

  • Ho, Thi Kieu Khanh;Kim, Inki;Jeon, Younghoon;Song, Jong-In;Gwak, Jeonghwan
    • Proceedings of the Korean Society of Computer Information Conference
    • /
    • 2021.07a
    • /
    • pp.305-307
    • /
    • 2021
  • Alzheimer's Disease (AD) is a cognitive disorder characterized by memory impairment that can be assessed at early stages based on administering clinical tests. However, the AD pathophysiological mechanism is still poorly understood due to the difficulty of distinguishing different levels of AD severity, even using a variety of brain modalities. Therefore, in this study, we present a hybrid EEG-fNIRS modalities to compensate for each other's weaknesses with the help of Machine Learning (ML) techniques for classifying four subject groups, including healthy controls (HC) and three distinguishable groups of AD levels. A concurrent EEF-fNIRS setup was used to record the data from 41 subjects during Oddball and 1-back tasks. We employed both a traditional neural network (NN) and a CNN-LSTM hybrid model for fNIRS and EEG, respectively. The final prediction was then obtained by using majority voting of those models. Classification results indicated that the hybrid EEG-fNIRS feature set achieved a higher accuracy (71.4%) by combining their complementary properties, compared to using EEG (67.9%) or fNIRS alone (68.9%). These findings demonstrate the potential of an EEG-fNIRS hybridization technique coupled with ML-based approaches for further AD studies.

  • PDF

Identification of Multiple Cancer Cell Lines from Microscopic Images via Deep Learning (심층 학습을 통한 암세포 광학영상 식별기법)

  • Park, Jinhyung;Choe, Se-woon
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2021.05a
    • /
    • pp.374-376
    • /
    • 2021
  • For the diagnosis of cancer-related diseases in clinical practice, pathological examination using biopsy is essential after basic diagnosis using imaging equipment. In order to proceed with such a biopsy, the assistance of an oncologist, clinical pathologist, etc. with specialized knowledge and the minimum required time are essential for confirmation. In recent years, research related to the establishment of a system capable of automatic classification of cancer cells using artificial intelligence is being actively conducted. However, previous studies show limitations in the type and accuracy of cells based on a limited algorithm. In this study, we propose a method to identify a total of 4 cancer cells through a convolutional neural network, a kind of deep learning. The optical images obtained through cell culture were learned through EfficientNet after performing pre-processing such as identification of the location of cells and image segmentation using OpenCV. The model used various hyper parameters based on EfficientNet, and trained InceptionV3 to compare and analyze the performance. As a result, cells were classified with a high accuracy of 96.8%, and this analysis method is expected to be helpful in confirming cancer.

  • PDF