• Title/Summary/Keyword: Convolutional Neural Networks (CNN)

Search Result 355, Processing Time 0.037 seconds

Compression of CNN Using Low-Rank Approximation and CP Decomposition Methods (저계수 행렬 근사 및 CP 분해 기법을 이용한 CNN 압축)

  • Moon, HyeonCheol;Moon, Gihwa;Kim, Jae-Gon
    • Journal of Broadcast Engineering
    • /
    • v.26 no.2
    • /
    • pp.125-131
    • /
    • 2021
  • In recent years, Convolutional Neural Networks (CNNs) have achieved outstanding performance in the fields of computer vision such as image classification, object detection, visual quality enhancement, etc. However, as huge amount of computation and memory are required in CNN models, there is a limitation in the application of CNN to low-power environments such as mobile or IoT devices. Therefore, the need for neural network compression to reduce the model size while keeping the task performance as much as possible has been emerging. In this paper, we propose a method to compress CNN models by combining matrix decomposition methods of LR (Low-Rank) approximation and CP (Canonical Polyadic) decomposition. Unlike conventional methods that apply one matrix decomposition method to CNN models, we selectively apply two decomposition methods depending on the layer types of CNN to enhance the compression performance. To evaluate the performance of the proposed method, we use the models for image classification such as VGG-16, RestNet50 and MobileNetV2 models. The experimental results show that the proposed method gives improved classification performance at the same range of 1.5 to 12.1 times compression ratio than the existing method that applies only the LR approximation.

A Tensor Space Model based Deep Neural Network for Automated Text Classification (자동문서분류를 위한 텐서공간모델 기반 심층 신경망)

  • Lim, Pu-reum;Kim, Han-joon
    • Database Research
    • /
    • v.34 no.3
    • /
    • pp.3-13
    • /
    • 2018
  • Text classification is one of the text mining technologies that classifies a given textual document into its appropriate categories and is used in various fields such as spam email detection, news classification, question answering, emotional analysis, and chat bot. In general, the text classification system utilizes machine learning algorithms, and among a number of algorithms, naïve Bayes and support vector machine, which are suitable for text data, are known to have reasonable performance. Recently, with the development of deep learning technology, several researches on applying deep neural networks such as recurrent neural networks (RNN) and convolutional neural networks (CNN) have been introduced to improve the performance of text classification system. However, the current text classification techniques have not yet reached the perfect level of text classification. This paper focuses on the fact that the text data is expressed as a vector only with the word dimensions, which impairs the semantic information inherent in the text, and proposes a neural network architecture based upon the semantic tensor space model.

Application of deep convolutional neural network for short-term precipitation forecasting using weather radar-based images

  • Le, Xuan-Hien;Jung, Sungho;Lee, Giha
    • Proceedings of the Korea Water Resources Association Conference
    • /
    • 2021.06a
    • /
    • pp.136-136
    • /
    • 2021
  • In this study, a deep convolutional neural network (DCNN) model is proposed for short-term precipitation forecasting using weather radar-based images. The DCNN model is a combination of convolutional neural networks, autoencoder neural networks, and U-net architecture. The weather radar-based image data used here are retrieved from competition for rainfall forecasting in Korea (AI Contest for Rainfall Prediction of Hydroelectric Dam Using Public Data), organized by Dacon under the sponsorship of the Korean Water Resources Association in October 2020. This data is collected from rainy events during the rainy season (April - October) from 2010 to 2017. These images have undergone a preprocessing step to convert from weather radar data to grayscale image data before they are exploited for the competition. Accordingly, each of these gray images covers a spatial dimension of 120×120 pixels and has a corresponding temporal resolution of 10 minutes. Here, each pixel corresponds to a grid of size 4km×4km. The DCNN model is designed in this study to provide 10-minute predictive images in advance. Then, precipitation information can be obtained from these forecast images through empirical conversion formulas. Model performance is assessed by comparing the Score index, which is defined based on the ratio of MAE (mean absolute error) to CSI (critical success index) values. The competition results have demonstrated the impressive performance of the DCNN model, where the Score value is 0.530 compared to the best value from the competition of 0.500, ranking 16th out of 463 participating teams. This study's findings exhibit the potential of applying the DCNN model to short-term rainfall prediction using weather radar-based images. As a result, this model can be applied to other areas with different spatiotemporal resolutions.

  • PDF

Improving Efficiency of Object Detection using Multiple Neural Networks (다중 신경망을 이용한 객체 탐지 효율성 개선방안)

  • Park, Dae-heum;Lim, Jong-hoon;Jang, Si-Woong
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2022.05a
    • /
    • pp.154-157
    • /
    • 2022
  • In the existing Tensorflow CNN environment, the object detection method is a method of performing object labeling and detection by Tensorflow itself. However, with the advent of YOLO, the efficiency of image object detection has increased. As a result, more deep layers can be built than existing neural networks, and the image object recognition rate can be increased. Therefore, in this paper, the detection ability and speed were compared and analyzed by designing an object detection system based on Darknet and YOLO and performing multi-layer construction and learning based on the existing convolutional neural network. For this reason, in this paper, a neural network methodology that efficiently uses Darknet's learning is presented.

  • PDF

SDCN: Synchronized Depthwise Separable Convolutional Neural Network for Single Image Super-Resolution

  • Muhammad, Wazir;Hussain, Ayaz;Shah, Syed Ali Raza;Shah, Jalal;Bhutto, Zuhaibuddin;Thaheem, Imdadullah;Ali, Shamshad;Masrour, Salman
    • International Journal of Computer Science & Network Security
    • /
    • v.21 no.11
    • /
    • pp.17-22
    • /
    • 2021
  • Recently, image super-resolution techniques used in convolutional neural networks (CNN) have led to remarkable performance in the research area of digital image processing applications and computer vision tasks. Convolutional layers stacked on top of each other can design a more complex network architecture, but they also use more memory in terms of the number of parameters and introduce the vanishing gradient problem during training. Furthermore, earlier approaches of single image super-resolution used interpolation technique as a pre-processing stage to upscale the low-resolution image into HR image. The design of these approaches is simple, but not effective and insert the newer unwanted pixels (noises) in the reconstructed HR image. In this paper, authors are propose a novel single image super-resolution architecture based on synchronized depthwise separable convolution with Dense Skip Connection Block (DSCB). In addition, unlike existing SR methods that only rely on single path, but our proposed method used the synchronizes path for generating the SISR image. Extensive quantitative and qualitative experiments show that our method (SDCN) achieves promising improvements than other state-of-the-art methods.

Senior Activity Recognition System using Time-series sensor data based on CNN-LSTM (CNN-LSTM 기반 시계열 센서 데이터를 이용한 노인 활동 인식 시스템)

  • Sunmin Lee;Nammee Moon
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2023.11a
    • /
    • pp.1230-1233
    • /
    • 2023
  • 최근, 65세 이상의 1인 가구가 급증함에 따라 노인을 대상으로 한 다양한 연구 및 서비스가 활발히 이루어지고 있다. 이에 본 논문에서는 시계열 센서 데이터를 이용하여 CNN-LSTM 기반의 노인 활동 인식 시스템을 제안한다. 수집된 데이터는 3축 가속도 센서가 내장된 2개의 디바이스를 등과 허벅지에 부착하였다. 수집 주기는 50hz로 진행되었으며, 각 행동은 2초를 기준으로 산정하였다. 학습데이터의 입력값으로 사용하기 위해, 슬라이딩 윈도우를 50%로 적용하여 시퀀스를 구성하였다. 모델은 특징을 반영하기 위한 CNN(Convolutional Neural Networks)과 시계열적 특성을 반영하기 위한 LSTM(Long-Short Term Memory)을 하이브리드한 1차원 형태의 CNN-LSTM 모델을 사용한다. 행동은 4가지로 분류하였으며, 97%의 정확도를 나타내고 있다.

Convolutional neural networks for automated tooth numbering on panoramic radiographs: A scoping review

  • Ramadhan Hardani Putra;Eha Renwi Astuti;Aga Satria Nurrachman;Dina Karimah Putri;Ahmad Badruddin Ghazali;Tjio Andrinanti Pradini;Dhinda Tiara Prabaningtyas
    • Imaging Science in Dentistry
    • /
    • v.53 no.4
    • /
    • pp.271-281
    • /
    • 2023
  • Purpose: The objective of this scoping review was to investigate the applicability and performance of various convolutional neural network (CNN) models in tooth numbering on panoramic radiographs, achieved through classification, detection, and segmentation tasks. Materials and Methods: An online search was performed of the PubMed, Science Direct, and Scopus databases. Based on the selection process, 12 studies were included in this review. Results: Eleven studies utilized a CNN model for detection tasks, 5 for classification tasks, and 3 for segmentation tasks in the context of tooth numbering on panoramic radiographs. Most of these studies revealed high performance of various CNN models in automating tooth numbering. However, several studies also highlighted limitations of CNNs, such as the presence of false positives and false negatives in identifying decayed teeth, teeth with crown prosthetics, teeth adjacent to edentulous areas, dental implants, root remnants, wisdom teeth, and root canal-treated teeth. These limitations can be overcome by ensuring both the quality and quantity of datasets, as well as optimizing the CNN architecture. Conclusion: CNNs have demonstrated high performance in automated tooth numbering on panoramic radiographs. Future development of CNN-based models for this purpose should also consider different stages of dentition, such as the primary and mixed dentition stages, as well as the presence of various tooth conditions. Ultimately, an optimized CNN architecture can serve as the foundation for an automated tooth numbering system and for further artificial intelligence research on panoramic radiographs for a variety of purposes.

Face Detection Using Multiple Filters and Hybrid Neural Networks (다중 필터와 복합형 신경망을 이용한 얼굴 검출 기법)

  • Cho, Il-Gook;Park, Hyun-Jung;Kim, Ho-Joon
    • Proceedings of the Korean Society of Broadcast Engineers Conference
    • /
    • 2005.11a
    • /
    • pp.191-194
    • /
    • 2005
  • 본 논문에서는 방송 영상에서 조명효과와 크기변화 등에 강인한 얼굴패턴 검출기법을 제시한다. 제안된 얼굴검출 모델은 영상 전처리 과정과 얼굴패턴 검출 과정으로 이루어진다. 전처리 과정은 조명변화에 대한 보정기능과 다중필터에 의한 후보영역 선별기능으로 구분된다. 얼굴패턴 검출과정은 다단계의 특징지도 생성과정과 패턴분류 과정으로 이루어진다. 특징지도를 생성하기 위하여 가보(Gabor) 필터계층을 포함하는 CNN(Convolutional Neural Networks)모델을 도입하였다. 다양한 배경을 고려한 효과적인 학습을 위하여 본 논문에서는 억제성의 뉴런(Inhibitory neuron)을 포함하는 구조의 CNN모델을 적용한다. CNN으로부터 추출되는 특징집합은 최종 단계에서 WFMM(Weighted Fuzzy Min Max) 모델을 사용하여 분류된다. 이때 사용되는 특징집합의 크기는 분류기의 규모 및 계산량의 결정적인 역할을 준다. 이에 본 연구에서는 최종 분류 과정에 사용되는 특징의 수를 효과적으로 줄이기 위해 FMM모델을 사용하는 적응적인 특징 선별 기법을 제안한다. 또한 실제 영상을 통한 실험결과로부터 제안된 이론의 타당성을 고찰한다.

  • PDF

Deep Learning based Color Restoration of Corrupted Black and White Facial Photos (딥러닝 기반 손상된 흑백 얼굴 사진 컬러 복원)

  • Woo, Shin Jae;Kim, Jong-Hyun;Lee, Jung;Song, Chang-Germ;Kim, Sun-Jeong
    • Journal of the Korea Computer Graphics Society
    • /
    • v.24 no.2
    • /
    • pp.1-9
    • /
    • 2018
  • In this paper, we propose a method to restore corrupted black and white facial images to color. Previous studies have shown that when coloring damaged black and white photographs, such as old ID photographs, the area around the damaged area is often incorrectly colored. To solve this problem, this paper proposes a method of restoring the damaged area of input photo first and then performing colorization based on the result. The proposed method consists of two steps: BEGAN (Boundary Equivalent Generative Adversarial Networks) model based restoration and CNN (Convolutional Neural Network) based coloring. Our method uses the BEGAN model, which enables a clearer and higher resolution image restoration than the existing methods using the DCGAN (Deep Convolutional Generative Adversarial Networks) model for image restoration, and performs colorization based on the restored black and white image. Finally, we confirmed that the experimental results of various types of facial images and masks can show realistic color restoration results in many cases compared with the previous studies.

Image-Based Automatic Detection of Construction Helmets Using R-FCN and Transfer Learning (R-FCN과 Transfer Learning 기법을 이용한 영상기반 건설 안전모 자동 탐지)

  • Park, Sangyoon;Yoon, Sanghyun;Heo, Joon
    • KSCE Journal of Civil and Environmental Engineering Research
    • /
    • v.39 no.3
    • /
    • pp.399-407
    • /
    • 2019
  • In Korea, the construction industry has been known to have the highest risk of safety accidents compared to other industries. Therefore, in order to improve safety in the construction industry, several researches have been carried out from the past. This study aims at improving safety of labors in construction site by constructing an effective automatic safety helmet detection system using object detection algorithm based on image data of construction field. Deep learning was conducted using Region-based Fully Convolutional Network (R-FCN) which is one of the object detection algorithms based on Convolutional Neural Network (CNN) with Transfer Learning technique. Learning was conducted with 1089 images including human and safety helmet collected from ImageNet and the mean Average Precision (mAP) of the human and the safety helmet was measured as 0.86 and 0.83, respectively.