• Title/Summary/Keyword: convolutional networks

Search Result 671, Processing Time 0.025 seconds

Training Artificial Neural Networks and Convolutional Neural Networks using WFSO Algorithm (WFSO 알고리즘을 이용한 인공 신경망과 합성곱 신경망의 학습)

  • Jang, Hyun-Woo;Jung, Sung Hoon
    • Journal of Digital Contents Society
    • /
    • v.18 no.5
    • /
    • pp.969-976
    • /
    • 2017
  • This paper proposes the learning method of an artificial neural network and a convolutional neural network using the WFSO algorithm developed as an optimization algorithm. Since the optimization algorithm searches based on a number of candidate solutions, it has a drawback in that it is generally slow, but it rarely falls into the local optimal solution and it is easy to parallelize. In addition, the artificial neural networks with non-differentiable activation functions can be trained and the structure and weights can be optimized at the same time. In this paper, we describe how to apply WFSO algorithm to artificial neural network learning and compare its performances with error back-propagation algorithm in multilayer artificial neural networks and convolutional neural networks.

A Study on Optimal Convolutional Neural Networks Backbone for Reinforced Concrete Damage Feature Extraction (철근콘크리트 손상 특성 추출을 위한 최적 컨볼루션 신경망 백본 연구)

  • Park, Younghoon
    • KSCE Journal of Civil and Environmental Engineering Research
    • /
    • v.43 no.4
    • /
    • pp.511-523
    • /
    • 2023
  • Research on the integration of unmanned aerial vehicles and deep learning for reinforced concrete damage detection is actively underway. Convolutional neural networks have a high impact on the performance of image classification, detection, and segmentation as backbones. The MobileNet, a pre-trained convolutional neural network, is efficient as a backbone for an unmanned aerial vehicle-based damage detection model because it can achieve sufficient accuracy with low computational complexity. Analyzing vanilla convolutional neural networks and MobileNet under various conditions, MobileNet was evaluated to have a verification accuracy 6.0~9.0% higher than vanilla convolutional neural networks with 15.9~22.9% lower computational complexity. MobileNetV2, MobileNetV3Large and MobileNetV3Small showed almost identical maximum verification accuracy, and the optimal conditions for MobileNet's reinforced concrete damage image feature extraction were analyzed to be the optimizer RMSprop, no dropout, and average pooling. The maximum validation accuracy of 75.49% for 7 types of damage detection based on MobilenetV2 derived in this study can be improved by image accumulation and continuous learning.

Efficient Fixed-Point Representation for ResNet-50 Convolutional Neural Network (ResNet-50 합성곱 신경망을 위한 고정 소수점 표현 방법)

  • Kang, Hyeong-Ju
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.22 no.1
    • /
    • pp.1-8
    • /
    • 2018
  • Recently, the convolutional neural network shows high performance in many computer vision tasks. However, convolutional neural networks require enormous amount of operation, so it is difficult to adopt them in the embedded environments. To solve this problem, many studies are performed on the ASIC or FPGA implementation, where an efficient representation method is required. The fixed-point representation is adequate for the ASIC or FPGA implementation but causes a performance degradation. This paper proposes a separate optimization of representations for the convolutional layers and the batch normalization layers. With the proposed method, the required bit width for the convolutional layers is reduced from 16 bits to 10 bits for the ResNet-50 neural network. Since the computation amount of the convolutional layers occupies the most of the entire computation, the bit width reduction in the convolutional layers enables the efficient implementation of the convolutional neural networks.

Face Anti-Spoofing Based on Combination of Luminance and Chrominance with Convolutional Neural Networks (합성곱 신경망 기반 밝기-색상 정보를 이용한 얼굴 위변조 검출 방법)

  • Kim, Eunseok;Kim, Wonjun
    • Journal of Broadcast Engineering
    • /
    • v.24 no.6
    • /
    • pp.1113-1121
    • /
    • 2019
  • In this paper, we propose the face anti-spoofing method based on combination of luminance and chrominance with convolutional neural networks. The proposed method extracts luminance and chrominance features independently from live and fake faces by using stacked convolutional neural networks and auxiliary networks. Unlike previous methods, an attention module has been adopted to adaptively combine extracted features instead of simply concatenating them. In addition, we propose a new loss function, called the contrast loss, to learn the classifier more efficiently. Specifically, the contrast loss improves the discriminative power of the features by maximizing the distance of the inter-class features while minimizing that of the intra-class features. Experimental results demonstrate that our method achieves the significant improvement for face anti-spoofing compared to existing methods.

Development of Deep Learning Models for Multi-class Sentiment Analysis (딥러닝 기반의 다범주 감성분석 모델 개발)

  • Syaekhoni, M. Alex;Seo, Sang Hyun;Kwon, Young S.
    • Journal of Information Technology Services
    • /
    • v.16 no.4
    • /
    • pp.149-160
    • /
    • 2017
  • Sentiment analysis is the process of determining whether a piece of document, text or conversation is positive, negative, neural or other emotion. Sentiment analysis has been applied for several real-world applications, such as chatbot. In the last five years, the practical use of the chatbot has been prevailing in many field of industry. In the chatbot applications, to recognize the user emotion, sentiment analysis must be performed in advance in order to understand the intent of speakers. The specific emotion is more than describing positive or negative sentences. In light of this context, we propose deep learning models for conducting multi-class sentiment analysis for identifying speaker's emotion which is categorized to be joy, fear, guilt, sad, shame, disgust, and anger. Thus, we develop convolutional neural network (CNN), long short term memory (LSTM), and multi-layer neural network models, as deep neural networks models, for detecting emotion in a sentence. In addition, word embedding process was also applied in our research. In our experiments, we have found that long short term memory (LSTM) model performs best compared to convolutional neural networks and multi-layer neural networks. Moreover, we also show the practical applicability of the deep learning models to the sentiment analysis for chatbot.

A Study of Active Pulse Classification Algorithm using Multi-label Convolutional Neural Networks (다중 레이블 콘볼루션 신경회로망을 이용한 능동펄스 식별 알고리즘 연구)

  • Kim, Guenhwan;Lee, Seokjin;Lee, Kyunkyung;Lee, Donghwa
    • Journal of Korea Society of Industrial Information Systems
    • /
    • v.25 no.4
    • /
    • pp.29-38
    • /
    • 2020
  • In this research, we proposed the active pulse classification algorithm using multi-label convolutional neural networks for active sonar system. The proposed algorithm has the advantage of being able to acquire the information of the active pulse at a time, unlike the existing single label-based algorithm, which has several neural network structures, and also has an advantage of simplifying the learning process. In order to verify the proposed algorithm, the neural network was trained using sea experimental data. As a result of the analysis, it was confirmed that the proposed algorithm converged, and through the analysis of the confusion matrix, it was confirmed that it has excellent active pulse classification performance.

Localization of ripe tomato bunch using deep neural networks and class activation mapping

  • Seung-Woo Kang;Soo-Hyun Cho;Dae-Hyun Lee;Kyung-Chul Kim
    • Korean Journal of Agricultural Science
    • /
    • v.50 no.3
    • /
    • pp.357-364
    • /
    • 2023
  • In this study, we propose a ripe tomato bunch localization method based on convolutional neural networks, to be applied in robotic harvesting systems. Tomato images were obtained from a smart greenhouse at the Rural Development Administration (RDA). The sample images for training were extracted based on tomato maturity and resized to 128 × 128 pixels for use in the classification model. The model was constructed based on four-layer convolutional neural networks, and the classes were determined based on stage of maturity, using a Softmax classifier. The localization of the ripe tomato bunch region was indicated on a class activation map. The class activation map could show the approximate location of the tomato bunch but tends to present a local part or a large part of the ripe tomato bunch region, which could lead to poor performance. Therefore, we suggest a recursive method to improve the performance of the model. The classification results indicated that the accuracy, precision, recall, and F1-score were 0.98, 0.87, 0.98, and 0.92, respectively. The localization performance was 0.52, estimated by the Intersection over Union (IoU), and through input recursion, the IoU was improved by 13%. Based on the results, the proposed localization of the ripe tomato bunch area can be incorporated in robotic harvesting systems to establish the optimal harvesting paths.

Correcting Misclassified Image Features with Convolutional Coding

  • Mun, Ye-Ji;Kim, Nayoung;Lee, Jieun;Kang, Je-Won
    • Proceedings of the Korean Society of Broadcast Engineers Conference
    • /
    • 2018.11a
    • /
    • pp.11-14
    • /
    • 2018
  • The aim of this study is to rectify the misclassified image features and enhance the performance of image classification tasks by incorporating a channel- coding technique, widely used in telecommunication. Specifically, the proposed algorithm employs the error - correcting mechanism of convolutional coding combined with the convolutional neural networks (CNNs) that are the state - of- the- arts image classifier s. We develop an encoder and a decoder to employ the error - correcting capability of the convolutional coding. In the encoder, the label values of the image data are converted to convolutional codes that are used as target outputs of the CNN, and the network is trained to minimize the Euclidean distance between the target output codes and the actual output codes. In order to correct misclassified features, the outputs of the network are decoded through the trellis structure with Viterbi algorithm before determining the final prediction. This paper demonstrates that the proposed architecture advances the performance of the neural networks compared to the traditional one- hot encoding method.

  • PDF

Implementation of Face Mask Detection (얼굴 마스크 탐지의 구현)

  • Park, Seong Hwan;Jung, Yuchul
    • Proceedings of the Korean Society of Computer Information Conference
    • /
    • 2021.01a
    • /
    • pp.17-19
    • /
    • 2021
  • 본 논문에서는 코로나19 사태에 대비하여 실시간으로 마스크를 제대로 쓴 사람과 제대로 쓰지 않은 사람을 구분하는 시스템을 제안한다. 이 시스템을 사용하기 위하여 모델 학습 시에 합성곱 신경망(CNN : Convolutional Neural Networks)를 사용한다. 학습된 모델을 토대로 영상에 적용 시 하르 특징 분류기(Haar Cascade Classifier)로 얼굴을 탐지하여 마스크 여부를 판단한다.

  • PDF

Bio-signal Data Augumentation Technique for CNN based Human Activity Recognition (CNN 기반 인간 동작 인식을 위한 생체신호 데이터의 증강 기법)

  • Gerelbat BatGerel;Chun-Ki Kwon
    • Journal of the Institute of Convergence Signal Processing
    • /
    • v.24 no.2
    • /
    • pp.90-96
    • /
    • 2023
  • Securing large amounts of training data in deep learning neural networks, including convolutional neural networks, is of importance for avoiding overfitting phenomenon or for the excellent performance. However, securing labeled training data in deep learning neural networks is very limited in reality. To overcome this, several augmentation methods have been proposed in the literature to generate an additional large amount of training data through transformation or manipulation of the already acquired traing data. However, unlike training data such as images and texts, it is barely to find an augmentation method in the literature that additionally generates bio-signal training data for convolutional neural network based human activity recognition. Thus, this study proposes a simple but effective augmentation method of bio-signal training data for convolutional neural network based human activity recognition. The usefulness of the proposed augmentation method is validated by showing that human activity is recognized with high accuracy by convolutional neural network trained with its augmented bio-signal training data.