• 제목/요약/키워드: Convolutional neural net

검색결과 299건 처리시간 0.027초

Convolutional Neural Networks for Character-level Classification

  • Ko, Dae-Gun;Song, Su-Han;Kang, Ki-Min;Han, Seong-Wook
    • IEIE Transactions on Smart Processing and Computing
    • /
    • 제6권1호
    • /
    • pp.53-59
    • /
    • 2017
  • Optical character recognition (OCR) automatically recognizes text in an image. OCR is still a challenging problem in computer vision. A successful solution to OCR has important device applications, such as text-to-speech conversion and automatic document classification. In this work, we analyze character recognition performance using the current state-of-the-art deep-learning structures. One is the AlexNet structure, another is the LeNet structure, and the other one is the SPNet structure. For this, we have built our own dataset that contains digits and upper- and lower-case characters. We experiment in the presence of salt-and-pepper noise or Gaussian noise, and report the performance comparison in terms of recognition error. Experimental results indicate by five-fold cross-validation that the SPNet structure (our approach) outperforms AlexNet and LeNet in recognition error.

DEXA에서 딥러닝 기반의 척골 및 요골 자동 분할 모델 (Automated Ulna and Radius Segmentation model based on Deep Learning on DEXA)

  • 김영재;박성진;김경래;김광기
    • 한국멀티미디어학회논문지
    • /
    • 제21권12호
    • /
    • pp.1407-1416
    • /
    • 2018
  • The purpose of this study was to train a model for the ulna and radius bone segmentation based on Convolutional Neural Networks and to verify the segmentation model. The data consisted of 840 training data, 210 tuning data, and 200 verification data. The learning model for the ulna and radius bone bwas based on U-Net (19 convolutional and 8 maximum pooling) and trained with 8 batch sizes, 0.0001 learning rate, and 200 epochs. As a result, the average sensitivity of the training data was 0.998, the specificity was 0.972, the accuracy was 0.979, and the Dice's similarity coefficient was 0.968. In the validation data, the average sensitivity was 0.961, specificity was 0.978, accuracy was 0.972, and Dice's similarity coefficient was 0.961. The performance of deep convolutional neural network based models for the segmentation was good for ulna and radius bone.

Performance Enhancement of Automatic Wood Classification of Korean Softwood by Ensembles of Convolutional Neural Networks

  • Kwon, Ohkyung;Lee, Hyung Gu;Yang, Sang-Yun;Kim, Hyunbin;Park, Se-Yeong;Choi, In-Gyu;Yeo, Hwanmyeong
    • Journal of the Korean Wood Science and Technology
    • /
    • 제47권3호
    • /
    • pp.265-276
    • /
    • 2019
  • In our previous study, the LeNet3 model successfully classified images from the transverse surfaces of five Korean softwood species (cedar, cypress, Korean pine, Korean red pine, and larch). However, a practical limitation exists in our system stemming from the nature of the training images obtained from the transverse plane of the wood species. In real-world applications, it is necessary to utilize images from the longitudinal surfaces of lumber. Thus, we improved our model by training it with images from the longitudinal and transverse surfaces of lumber. Because the longitudinal surface has complex but less distinguishable features than the transverse surface, the classification performance of the LeNet3 model decreases when we include images from the longitudinal surfaces of the five Korean softwood species. To remedy this situation, we adopt ensemble methods that can enhance the classification performance. Herein, we investigated the use of ensemble models from the LeNet and MiniVGGNet models to automatically classify the transverse and longitudinal surfaces of the five Korean softwoods. Experimentally, the best classification performance was achieved via an ensemble model comprising the LeNet2, LeNet3, and MiniVGGNet4 models trained using input images of $128{\times}128{\times}3pixels$ via the averaging method. The ensemble model showed an F1 score greater than 0.98. The classification performance for the longitudinal surfaces of Korean pine and Korean red pine was significantly improved by the ensemble model compared to individual convolutional neural network models such as LeNet3.

A ResNet based multiscale feature extraction for classifying multi-variate medical time series

  • Zhu, Junke;Sun, Le;Wang, Yilin;Subramani, Sudha;Peng, Dandan;Nicolas, Shangwe Charmant
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • 제16권5호
    • /
    • pp.1431-1445
    • /
    • 2022
  • We construct a deep neural network model named ECGResNet. This model can diagnosis diseases based on 12-lead ECG data of eight common cardiovascular diseases with a high accuracy. We chose the 16 Blocks of ResNet50 as the main body of the model and added the Squeeze-and-Excitation module to learn the data information between channels adaptively. We modified the first convolutional layer of ResNet50 which has a convolutional kernel of 7 to a superposition of convolutional kernels of 8 and 16 as our feature extraction method. This way allows the model to focus on the overall trend of the ECG signal while also noticing subtle changes. The model further improves the accuracy of cardiovascular and cerebrovascular disease classification by using a fully connected layer that integrates factors such as gender and age. The ECGResNet model adds Dropout layers to both the residual block and SE module of ResNet50, further avoiding the phenomenon of model overfitting. The model was eventually trained using a five-fold cross-validation and Flooding training method, with an accuracy of 95% on the test set and an F1-score of 0.841.We design a new deep neural network, innovate a multi-scale feature extraction method, and apply the SE module to extract features of ECG data.

Effective Hand Gesture Recognition by Key Frame Selection and 3D Neural Network

  • Hoang, Nguyen Ngoc;Lee, Guee-Sang;Kim, Soo-Hyung;Yang, Hyung-Jeong
    • 스마트미디어저널
    • /
    • 제9권1호
    • /
    • pp.23-29
    • /
    • 2020
  • This paper presents an approach for dynamic hand gesture recognition by using algorithm based on 3D Convolutional Neural Network (3D_CNN), which is later extended to 3D Residual Networks (3D_ResNet), and the neural network based key frame selection. Typically, 3D deep neural network is used to classify gestures from the input of image frames, randomly sampled from a video data. In this work, to improve the classification performance, we employ key frames which represent the overall video, as the input of the classification network. The key frames are extracted by SegNet instead of conventional clustering algorithms for video summarization (VSUMM) which require heavy computation. By using a deep neural network, key frame selection can be performed in a real-time system. Experiments are conducted using 3D convolutional kernels such as 3D_CNN, Inflated 3D_CNN (I3D) and 3D_ResNet for gesture classification. Our algorithm achieved up to 97.8% of classification accuracy on the Cambridge gesture dataset. The experimental results show that the proposed approach is efficient and outperforms existing methods.

비전 트랜스포머 인코더가 포함된 U-net을 이용한 대장 내시경 이미지의 폴립 분할 (U-net with vision transformer encoder for polyp segmentation in colonoscopy images)

  • 겔란 아야나;최세운
    • 한국정보통신학회:학술대회논문집
    • /
    • 한국정보통신학회 2022년도 추계학술대회
    • /
    • pp.97-99
    • /
    • 2022
  • 대장암의 조기 발견과 치료를 위해서는 정확한 폴립의 분할이 중요하나 다음과 같은 제약이 따른다. 개별 폴립의 위치, 크기 및 모양이 서로 상이하며, 모션 흐림 및 빛 반사와 같은 특정 상황에서 폴립과 주변 환경 간에 상당한 정도의 유사성이 존재한다. 인코더와 디코더 역할을 하는 Convolutional Neural Networks로 구성된 U-net은 이러한 한계를 극복하기 위해 다양하게 사용된다. 본 연구는 보다 정확한 폴립 분할을 위한 비전트랜스포머가 포함된 U-net 아키텍처를 제안하였고, 그 결과 제안된 방식은 표준 U-net 아키텍처보다 더 나은 성능을 보였음을 확인할 수 있었다.

  • PDF

Explainable radionuclide identification algorithm based on the convolutional neural network and class activation mapping

  • Yu Wang;Qingxu Yao;Quanhu Zhang;He Zhang;Yunfeng Lu;Qimeng Fan;Nan Jiang;Wangtao Yu
    • Nuclear Engineering and Technology
    • /
    • 제54권12호
    • /
    • pp.4684-4692
    • /
    • 2022
  • Radionuclide identification is an important part of the nuclear material identification system. The development of artificial intelligence and machine learning has made nuclide identification rapid and automatic. However, many methods directly use existing deep learning models to analyze the gamma-ray spectrum, which lacks interpretability for researchers. This study proposes an explainable radionuclide identification algorithm based on the convolutional neural network and class activation mapping. This method shows the area of interest of the neural network on the gamma-ray spectrum by generating a class activation map. We analyzed the class activation map of the gamma-ray spectrum of different types, different gross counts, and different signal-to-noise ratios. The results show that the convolutional neural network attempted to learn the relationship between the input gamma-ray spectrum and the nuclide type, and could identify the nuclide based on the photoelectric peak and Compton edge. Furthermore, the results explain why the neural network could identify gamma-ray spectra with low counts and low signal-to-noise ratios. Thus, the findings improve researchers' confidence in the ability of neural networks to identify nuclides and promote the application of artificial intelligence methods in the field of nuclide identification.

Deep Compression의 프루닝 문턱값 동적 조정 (Dynamic Adjustment of the Pruning Threshold in Deep Compression)

  • 이여진;박한훈
    • 융합신호처리학회논문지
    • /
    • 제22권3호
    • /
    • pp.99-103
    • /
    • 2021
  • 최근 CNN(Convolutional Neural Network)이 다양한 컴퓨터 비전 분야에서 우수한 성능으로 널리 사용되고 있다. 그러나 CNN은 계산 집약적이고 많은 메모리가 요구되어 한정적인 하드웨어 자원을 가지는 모바일이나 IoT(Internet of Things) 기기에 적용하기 어렵다. 이런 한계를 해결하기 위해, 기존의 학습된 모델의 성능을 최대한 유지하며 네트워크의 크기를 줄이는 인공신경망 경량화 연구가 진행되고 있다. 본 논문은 신경망 압축 기술 중 하나인 프루닝(Pruning)의 문턱값을 동적으로 조정하는 CNN 압축 기법을 제안한다. 프루닝될 가중치를 결정하는 문턱값을 실험적, 경험적으로 정하는 기존의 기술과 달리 정확도의 저하를 방지하는 최적의 문턱값을 동적으로 찾을 수 있으며, 경량화된 신경망을 얻는 시간을 단축할 수 있다. 제안 기법의 성능 검증을 위해 MNIST 데이터 셋을 사용하여 LeNet을 훈련시켰으며, 정확도 손실 없이 약 1.3 ~ 3배의 시간을 단축하여 경량화된 LeNet을 얻을 수 있었다.

Transfer learning for crack detection in concrete structures: Evaluation of four models

  • Ali Bagheri;Mohammadreza Mosalmanyazdi;Hasanali Mosalmanyazdi
    • Structural Engineering and Mechanics
    • /
    • 제91권2호
    • /
    • pp.163-175
    • /
    • 2024
  • The objective of this research is to improve public safety in civil engineering by recognizing fractures in concrete structures quickly and correctly. The study offers a new crack detection method based on advanced image processing and machine learning techniques, specifically transfer learning with convolutional neural networks (CNNs). Four pre-trained models (VGG16, AlexNet, ResNet18, and DenseNet161) were fine-tuned to detect fractures in concrete surfaces. These models constantly produced accuracy rates greater than 80%, showing their ability to automate fracture identification and potentially reduce structural failure costs. Furthermore, the study expands its scope beyond crack detection to identify concrete health, using a dataset with a wide range of surface defects and anomalies including cracks. Notably, using VGG16, which was chosen as the most effective network architecture from the first phase, the study achieves excellent accuracy in classifying concrete health, demonstrating the model's satisfactorily performance even in more complex scenarios.

A review and comparison of convolution neural network models under a unified framework

  • Park, Jimin;Jung, Yoonsuh
    • Communications for Statistical Applications and Methods
    • /
    • 제29권2호
    • /
    • pp.161-176
    • /
    • 2022
  • There has been active research in image classification using deep learning convolutional neural network (CNN) models. ImageNet large-scale visual recognition challenge (ILSVRC) (2010-2017) was one of the most important competitions that boosted the development of efficient deep learning algorithms. This paper introduces and compares six monumental models that achieved high prediction accuracy in ILSVRC. First, we provide a review of the models to illustrate their unique structure and characteristics of the models. We then compare those models under a unified framework. For this reason, additional devices that are not crucial to the structure are excluded. Four popular data sets with different characteristics are then considered to measure the prediction accuracy. By investigating the characteristics of the data sets and the models being compared, we provide some insight into the architectural features of the models.