• Title/Summary/Keyword: convolution network

Search Result 521, Processing Time 0.032 seconds

Artificial Intelligence Algorithms for Identification of Handwriting (효과적인 필기체 인식을 위한 인공지능 알고리즘)

  • Kim, Seung-Ju;Lee, Jae-Yung;Han, Jong-Ki
    • Proceedings of the Korean Society of Broadcast Engineers Conference
    • /
    • 2016.11a
    • /
    • pp.151-153
    • /
    • 2016
  • 최근 스마트폰, PC, 태블릿 같은 전자기기들이 발전하면서 기계를 통해 소통하는 시대가 왔다. 기계와 소통하기 위해 우리가 사용하는 문자를 인식하는 것은 중요한 일이다. 이런 전자기기들이 문자, 영상인식을 해야 할 필요성이 더욱 증가함에 따라 머신러닝의 중요성이 대두되었다. 머신러닝은 컴퓨터의 학습을 위해 알고리즘과 기술을 개발하는 분야를 말한다. 머신러닝의 기법과 관련된 알고리즘의 종류는 수없이 많다. 그 중에서도 Neural Network는 사람의 뇌 신경구조를 토대로 착안하여 네트워크를 만들고 이를 학습에 이용한 머신러닝 기법이다. 이런 인공지능 알고리즘인 Neural Network 구조를 바탕으로 특징을 추출하여 학습을 하는 Convolution Neural Network 기법의 사용이 늘고 있다. 본 논문에서는 Neural Network와 Convolution Neural Network의 알고리즘을 이용한 필기체 인식 실험을 하고 그 내용을 비교하였다.

  • PDF

A Recommendation Model based on Character-level Deep Convolution Neural Network (문자 수준 딥 컨볼루션 신경망 기반 추천 모델)

  • Ji, JiaQi;Chung, Yeongjee
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.23 no.3
    • /
    • pp.237-246
    • /
    • 2019
  • In order to improve the accuracy of the rating prediction of the recommendation model, not only user-item rating data are used but also consider auxiliary information of item such as comments, tags, or descriptions. The traditional approaches use a word-level model of the bag-of-words for the auxiliary information. This model, however, cannot utilize the auxiliary information effectively, which leads to shallow understanding of auxiliary information. Convolution neural network (CNN) can capture and extract feature vector from auxiliary information effectively. Thus, this paper proposes character-level deep-Convolution Neural Network based matrix factorization (Char-DCNN-MF) that integrates deep CNN into matrix factorization for a novel recommendation model. Char-DCNN-MF can deeper understand auxiliary information and further enhance recommendation performance. Experiments are performed on three different real data sets, and the results show that Char-DCNN-MF performs significantly better than other comparative models.

A Study on the Analysis of Jeju Island Precipitation Patterns using the Convolution Neural Network (합성곱신경망을 이용한 제주도 강수패턴 분석 연구)

  • Lee, Dong-Hoon;Lee, Bong-Kyu
    • Journal of Software Assessment and Valuation
    • /
    • v.15 no.2
    • /
    • pp.59-66
    • /
    • 2019
  • Since Jeju is the absolute weight of agriculture and tourism, the analysis of precipitation is more important than other regions. Currently, some numerical models are used for analysis of precipitation of Jeju Island using observation data from meteorological satellites. However, since precipitation changes are more diverse than other regions, it is difficult to obtain satisfactory results using the existing numerical models. In this paper, we propose a Jeju precipitation pattern analysis method using the texture analysis method based on Convolution Neural Network (CNN). The proposed method converts the water vapor image and the temperature information of the area of ​​Jeju Island from the weather satellite into texture images. Then converted images are fed into the CNN to analyse the precipitation patterns of Jeju Island. We implement the proposed method and show the effectiveness of the proposed method through experiments.

Optimization of Deep Learning Model Using Genetic Algorithm in PET-CT Image Alzheimer's Classification (PET-CT 영상 알츠하이머 분류에서 유전 알고리즘 이용한 심층학습 모델 최적화)

  • Lee, Sanghyeop;Kang, Do-Young;Song, Jongkwan;Park, Jangsik
    • Journal of Korea Multimedia Society
    • /
    • v.23 no.9
    • /
    • pp.1129-1138
    • /
    • 2020
  • The performance of convolutional deep learning networks is generally determined according to parameters of target dataset, structure of network, convolution kernel, activation function, and optimization algorithm. In this paper, a genetic algorithm is used to select the appropriate deep learning model and parameters for Alzheimer's classification and to compare the learning results with preliminary experiment. We compare and analyze the Alzheimer's disease classification performance of VGG-16, GoogLeNet, and ResNet to select an effective network for detecting AD and MCI. The simulation results show that the network structure is ResNet, the activation function is ReLU, the optimization algorithm is Adam, and the convolution kernel has a 3-dilated convolution filter for the accuracy of dementia medical images.

A Study on the Analysis of Structural Textures using CNN (Convolution Neural Network) (합성곱신경망을 이용한 구조적 텍스처 분석연구)

  • Lee, Bongkyu
    • The Journal of the Institute of Internet, Broadcasting and Communication
    • /
    • v.20 no.4
    • /
    • pp.201-205
    • /
    • 2020
  • The structural texture is defined as a form which a texel is regularly repeated in the texture. Structural texture analysis/recognition has various industrial applications, such as automatic inspection of textiles, automatic testing of metal surfaces, and automatic analysis of micro images. In this paper, we propose a Convolution Neural Network (CNN) based system for structural texture analysis. The proposed method learns texles, which are components of textures to be classified. Then, this trained CNN recognizes a structural texture using a partial image obtained from input texture. The experiment shows the superiority of the proposed system.

Security Vulnerability Verification for Open Deep Learning Libraries (공개 딥러닝 라이브러리에 대한 보안 취약성 검증)

  • Jeong, JaeHan;Shon, Taeshik
    • Journal of the Korea Institute of Information Security & Cryptology
    • /
    • v.29 no.1
    • /
    • pp.117-125
    • /
    • 2019
  • Deep Learning, which is being used in various fields recently, is being threatened with Adversarial Attack. In this paper, we experimentally verify that the classification accuracy is lowered by adversarial samples generated by malicious attackers in image classification models. We used MNIST dataset and measured the detection accuracy by injecting adversarial samples into the Autoencoder classification model and the CNN (Convolution neural network) classification model, which are created using the Tensorflow library and the Pytorch library. Adversarial samples were generated by transforming MNIST test dataset with JSMA(Jacobian-based Saliency Map Attack) and FGSM(Fast Gradient Sign Method). When injected into the classification model, detection accuracy decreased by at least 21.82% up to 39.08%.

A Study on Improving Speed of Interesting Region Detection Based on Fully Convolutional Network (Fully Convolutional Network 기반 관심 영역 검출 기법의 속도 개선 연구)

  • Hwang, Hyun-Su;Jung, Jin-woo;Kim, Yong-Hwan;Choe, Yoon-Sik
    • Proceedings of the Korean Society of Broadcast Engineers Conference
    • /
    • 2018.06a
    • /
    • pp.322-325
    • /
    • 2018
  • 영상의 관심 영역 검출은 영상처리 및 컴퓨터 비전 응용 분야에서 꾸준하게 사용되고 있는 기법이다. 특히, 근래 심층신경망 연구의 급격한 발전에 힘입어 심층신경망을 이용한 관심 영역 검출 기법에 대한 연구가 활발하게 진행되고 있다. 한편 Fully Convolutional Network(이하 FCN)은 본래 심층 예측(Dense Prediction)을 통한 의미론적 영상 분할(Semantic Segmentation)을 수행하기 위해 제안된 심층신경망 구조이다. FCN을 영상의 관심 영역 검출에 활용하여도 기존 관심 영역 검출 기법과 비교하여 충분히 좋은 성능을 발휘할 수 있다. 그러나 FCN에 사용되는 convolution 층의 수가 많고, 이에 따른 가중치(weight)의 개수도 기하급수적으로 늘어나 검출에 필요한 시간 복잡도가 매우 크다는 문제점이 있다. 따라서 본 논문에서는 기존 FCN이 가진 검출 시간 복잡도의 문제점을 convolution 층의 가중치 관점에서 해결하고자 이를 조절하여 FCN의 관심 영역 검출 속도를 향상시키는 방법을 제안한다. 적절한 convolution 층의 가중치를 조절함으로써, MSRA10K 데이터셋 환경에서 검출 정확도를 크게 저하시키지 않고도 최대 약 20.5%만큼 검출 속도를 향상시킬 수 있었다.

  • PDF

Pixel-based crack image segmentation in steel structures using atrous separable convolution neural network

  • Ta, Quoc-Bao;Pham, Quang-Quang;Kim, Yoon-Chul;Kam, Hyeon-Dong;Kim, Jeong-Tae
    • Structural Monitoring and Maintenance
    • /
    • v.9 no.3
    • /
    • pp.289-303
    • /
    • 2022
  • In this study, the impact of assigned pixel labels on the accuracy of crack image identification of steel structures is examined by using an atrous separable convolution neural network (ASCNN). Firstly, images containing fatigue cracks collected from steel structures are classified into four datasets by assigning different pixel labels based on image features. Secondly, the DeepLab v3+ algorithm is used to determine optimal parameters of the ASCNN model by maximizing the average mean-intersection-over-union (mIoU) metric of the datasets. Thirdly, the ASCNN model is trained for various image sizes and hyper-parameters, such as the learning rule, learning rate, and epoch. The optimal parameters of the ASCNN model are determined based on the average mIoU metric. Finally, the trained ASCNN model is evaluated by using 10% untrained images. The result shows that the ASCNN model can segment cracks and other objects in the captured images with an average mIoU of 0.716.

Semantic crack-image identification framework for steel structures using atrous convolution-based Deeplabv3+ Network

  • Ta, Quoc-Bao;Dang, Ngoc-Loi;Kim, Yoon-Chul;Kam, Hyeon-Dong;Kim, Jeong-Tae
    • Smart Structures and Systems
    • /
    • v.30 no.1
    • /
    • pp.17-34
    • /
    • 2022
  • For steel structures, fatigue cracks are critical damage induced by long-term cycle loading and distortion effects. Vision-based crack detection can be a solution to ensure structural integrity and performance by continuous monitoring and non-destructive assessment. A critical issue is to distinguish cracks from other features in captured images which possibly consist of complex backgrounds such as handwritings and marks, which were made to record crack patterns and lengths during periodic visual inspections. This study presents a parametric study on image-based crack identification for orthotropic steel bridge decks using captured images with complicated backgrounds. Firstly, a framework for vision-based crack segmentation using the atrous convolution-based Deeplapv3+ network (ACDN) is designed. Secondly, features on crack images are labeled to build three databanks by consideration of objects in the backgrounds. Thirdly, evaluation metrics computed from the trained ACDN models are utilized to evaluate the effects of obstacles on crack detection results. Finally, various training parameters, including image sizes, hyper-parameters, and the number of training images, are optimized for the ACDN model of crack detection. The result demonstrated that fatigue cracks could be identified by the trained ACDN models, and the accuracy of the crack-detection result was improved by optimizing the training parameters. It enables the applicability of the vision-based technique for early detecting tiny fatigue cracks in steel structures.

MLSE-Net: Multi-level Semantic Enriched Network for Medical Image Segmentation

  • Di Gai;Heng Luo;Jing He;Pengxiang Su;Zheng Huang;Song Zhang;Zhijun Tu
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.17 no.9
    • /
    • pp.2458-2482
    • /
    • 2023
  • Medical image segmentation techniques based on convolution neural networks indulge in feature extraction triggering redundancy of parameters and unsatisfactory target localization, which outcomes in less accurate segmentation results to assist doctors in diagnosis. In this paper, we propose a multi-level semantic-rich encoding-decoding network, which consists of a Pooling-Conv-Former (PCFormer) module and a Cbam-Dilated-Transformer (CDT) module. In the PCFormer module, it is used to tackle the issue of parameter explosion in the conservative transformer and to compensate for the feature loss in the down-sampling process. In the CDT module, the Cbam attention module is adopted to highlight the feature regions by blending the intersection of attention mechanisms implicitly, and the Dilated convolution-Concat (DCC) module is designed as a parallel concatenation of multiple atrous convolution blocks to display the expanded perceptual field explicitly. In addition, MultiHead Attention-DwConv-Transformer (MDTransformer) module is utilized to evidently distinguish the target region from the background region. Extensive experiments on medical image segmentation from Glas, SIIM-ACR, ISIC and LGG demonstrated that our proposed network outperforms existing advanced methods in terms of both objective evaluation and subjective visual performance.