• Title/Summary/Keyword: Convolutional neural networks

Search Result 619, Processing Time 0.028 seconds

Classification of Warhead and Debris using CFAR and Convolutional Neural Networks (CFAR와 합성곱 신경망을 이용한 기두부와 단 분리 시 조각 구분)

  • Seol, Seung-Hwan;Choi, In-Sik
    • The Journal of Korean Institute of Information Technology
    • /
    • v.17 no.6
    • /
    • pp.85-94
    • /
    • 2019
  • Warhead and debris show the different micro-Doppler frequency shape in the spectrogram because of the different micro motion. So we can classify them using the micro-Doppler features. In this paper, we classified warhead and debris in the separation phase using CNN(Convolutional Neural Networks). For the input image of CNN, we used micro-Doppler spectrogram. In addition, to improve classification performance of warhead and debris, we applied the preprocessing using CA-CFAR to the micro-Doppler spectrogram. As a result, when the preprocessing of micro-Doppler spectrogram was used, classification performance is improved in all signal-to-noise ratio(SNR).

Extraction of Protein-Protein Interactions based on Convolutional Neural Network (CNN) (Convolutional Neural Network (CNN) 기반의 단백질 간 상호 작용 추출)

  • Choi, Sung-Pil
    • KIISE Transactions on Computing Practices
    • /
    • v.23 no.3
    • /
    • pp.194-198
    • /
    • 2017
  • In this paper, we propose a revised Deep Convolutional Neural Network (DCNN) model to extract Protein-Protein Interaction (PPIs) from the scientific literature. The proposed method has the merit of improving performance by applying various global features in addition to the simple lexical features used in conventional relation extraction approaches. In the experiments using AIMed, which is the most famous collection used for PPI extraction, the proposed model shows state-of-the art scores (78.0 F-score) revealing the best performance so far in this domain. Also, the paper shows that, without conducting feature engineering using complicated language processing, convolutional neural networks with embedding can achieve superior PPIE performance.

A study in Hangul font characteristics using convolutional neural networks (컨볼루션 뉴럴 네트워크를 이용한 한글 서체 특징 연구)

  • Hwang, In-Kyeong;Won, Joong-Ho
    • The Korean Journal of Applied Statistics
    • /
    • v.32 no.4
    • /
    • pp.573-591
    • /
    • 2019
  • Classification criteria for Korean alphabet (Hangul) fonts are undeveloped in comparison to numerical classification systems for Roman alphabet fonts. This study finds important features that distinguish typeface styles in order to help develop numerical criteria for Hangul font classification. We find features that determine the characteristics of the two different styles using a convolutional neural network to create a model that analyzes the learned filters as well as distinguishes between serif and sans-serif styles.

Efficient Fixed-Point Representation for ResNet-50 Convolutional Neural Network (ResNet-50 합성곱 신경망을 위한 고정 소수점 표현 방법)

  • Kang, Hyeong-Ju
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.22 no.1
    • /
    • pp.1-8
    • /
    • 2018
  • Recently, the convolutional neural network shows high performance in many computer vision tasks. However, convolutional neural networks require enormous amount of operation, so it is difficult to adopt them in the embedded environments. To solve this problem, many studies are performed on the ASIC or FPGA implementation, where an efficient representation method is required. The fixed-point representation is adequate for the ASIC or FPGA implementation but causes a performance degradation. This paper proposes a separate optimization of representations for the convolutional layers and the batch normalization layers. With the proposed method, the required bit width for the convolutional layers is reduced from 16 bits to 10 bits for the ResNet-50 neural network. Since the computation amount of the convolutional layers occupies the most of the entire computation, the bit width reduction in the convolutional layers enables the efficient implementation of the convolutional neural networks.

A Study of Active Pulse Classification Algorithm using Multi-label Convolutional Neural Networks (다중 레이블 콘볼루션 신경회로망을 이용한 능동펄스 식별 알고리즘 연구)

  • Kim, Guenhwan;Lee, Seokjin;Lee, Kyunkyung;Lee, Donghwa
    • Journal of Korea Society of Industrial Information Systems
    • /
    • v.25 no.4
    • /
    • pp.29-38
    • /
    • 2020
  • In this research, we proposed the active pulse classification algorithm using multi-label convolutional neural networks for active sonar system. The proposed algorithm has the advantage of being able to acquire the information of the active pulse at a time, unlike the existing single label-based algorithm, which has several neural network structures, and also has an advantage of simplifying the learning process. In order to verify the proposed algorithm, the neural network was trained using sea experimental data. As a result of the analysis, it was confirmed that the proposed algorithm converged, and through the analysis of the confusion matrix, it was confirmed that it has excellent active pulse classification performance.

A Study on Optimal Convolutional Neural Networks Backbone for Reinforced Concrete Damage Feature Extraction (철근콘크리트 손상 특성 추출을 위한 최적 컨볼루션 신경망 백본 연구)

  • Park, Younghoon
    • KSCE Journal of Civil and Environmental Engineering Research
    • /
    • v.43 no.4
    • /
    • pp.511-523
    • /
    • 2023
  • Research on the integration of unmanned aerial vehicles and deep learning for reinforced concrete damage detection is actively underway. Convolutional neural networks have a high impact on the performance of image classification, detection, and segmentation as backbones. The MobileNet, a pre-trained convolutional neural network, is efficient as a backbone for an unmanned aerial vehicle-based damage detection model because it can achieve sufficient accuracy with low computational complexity. Analyzing vanilla convolutional neural networks and MobileNet under various conditions, MobileNet was evaluated to have a verification accuracy 6.0~9.0% higher than vanilla convolutional neural networks with 15.9~22.9% lower computational complexity. MobileNetV2, MobileNetV3Large and MobileNetV3Small showed almost identical maximum verification accuracy, and the optimal conditions for MobileNet's reinforced concrete damage image feature extraction were analyzed to be the optimizer RMSprop, no dropout, and average pooling. The maximum validation accuracy of 75.49% for 7 types of damage detection based on MobilenetV2 derived in this study can be improved by image accumulation and continuous learning.

Video Expression Recognition Method Based on Spatiotemporal Recurrent Neural Network and Feature Fusion

  • Zhou, Xuan
    • Journal of Information Processing Systems
    • /
    • v.17 no.2
    • /
    • pp.337-351
    • /
    • 2021
  • Automatically recognizing facial expressions in video sequences is a challenging task because there is little direct correlation between facial features and subjective emotions in video. To overcome the problem, a video facial expression recognition method using spatiotemporal recurrent neural network and feature fusion is proposed. Firstly, the video is preprocessed. Then, the double-layer cascade structure is used to detect a face in a video image. In addition, two deep convolutional neural networks are used to extract the time-domain and airspace facial features in the video. The spatial convolutional neural network is used to extract the spatial information features from each frame of the static expression images in the video. The temporal convolutional neural network is used to extract the dynamic information features from the optical flow information from multiple frames of expression images in the video. A multiplication fusion is performed with the spatiotemporal features learned by the two deep convolutional neural networks. Finally, the fused features are input to the support vector machine to realize the facial expression classification task. The experimental results on cNTERFACE, RML, and AFEW6.0 datasets show that the recognition rates obtained by the proposed method are as high as 88.67%, 70.32%, and 63.84%, respectively. Comparative experiments show that the proposed method obtains higher recognition accuracy than other recently reported methods.

Active Sonar Target/Non-target Classification using Convolutional Neural Networks (CNN을 이용한 능동 소나 표적/비표적 분류)

  • Kim, Dongwook;Seok, Jongwon;Bae, Keunsung
    • Journal of Korea Multimedia Society
    • /
    • v.21 no.9
    • /
    • pp.1062-1067
    • /
    • 2018
  • Conventional active sonar technology has relied heavily on the hearing of sonar operator, but recently, many techniques for automatic detection and classification have been studied. In this paper, we extract the image data from the spectrogram of the active sonar signal and classify the extracted data using CNN(convolutional neural networks), which has recently presented excellent performance improvement in the field of pattern recognition. First, we divided entire data set into eight classes depending on the ratio containing the target. Then, experiments were conducted to classify the eight classes data using proposed CNN structure, and the results were analyzed.

3D Convolutional Neural Networks based Fall Detection with Thermal Camera (열화상 카메라를 이용한 3D 컨볼루션 신경망 기반 낙상 인식)

  • Kim, Dae-Eon;Jeon, BongKyu;Kwon, Dong-Soo
    • The Journal of Korea Robotics Society
    • /
    • v.13 no.1
    • /
    • pp.45-54
    • /
    • 2018
  • This paper presents a vision-based fall detection system to automatically monitor and detect people's fall accidents, particularly those of elderly people or patients. For video analysis, the system should be able to extract both spatial and temporal features so that the model captures appearance and motion information simultaneously. Our approach is based on 3-dimensional convolutional neural networks, which can learn spatiotemporal features. In addition, we adopts a thermal camera in order to handle several issues regarding usability, day and night surveillance and privacy concerns. We design a pan-tilt camera with two actuators to extend the range of view. Performance is evaluated on our thermal dataset: TCL Fall Detection Dataset. The proposed model achieves 90.2% average clip accuracy which is better than other approaches.

POI Recommendation Method Based on Multi-Source Information Fusion Using Deep Learning in Location-Based Social Networks

  • Sun, Liqiang
    • Journal of Information Processing Systems
    • /
    • v.17 no.2
    • /
    • pp.352-368
    • /
    • 2021
  • Sign-in point of interest (POI) are extremely sparse in location-based social networks, hindering recommendation systems from capturing users' deep-level preferences. To solve this problem, we propose a content-aware POI recommendation algorithm based on a convolutional neural network. First, using convolutional neural networks to process comment text information, we model location POI and user latent factors. Subsequently, the objective function is constructed by fusing users' geographical information and obtaining the emotional category information. In addition, the objective function comprises matrix decomposition and maximisation of the probability objective function. Finally, we solve the objective function efficiently. The prediction rate and F1 value on the Instagram-NewYork dataset are 78.32% and 76.37%, respectively, and those on the Instagram-Chicago dataset are 85.16% and 83.29%, respectively. Comparative experiments show that the proposed method can obtain a higher precision rate than several other newer recommended methods.