• Title/Summary/Keyword: Convolutional Neural Network

Search Result 1,535, Processing Time 0.022 seconds

Application of CNN for Fish Species Classification (어종 분류를 위한 CNN의 적용)

  • Park, Jin-Hyun;Hwang, Kwang-Bok;Park, Hee-Mun;Choi, Young-Kiu
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.23 no.1
    • /
    • pp.39-46
    • /
    • 2019
  • In this study, before system development for the elimination of foreign fish species, we propose an algorithm to classify fish species by training fish images with CNN. The raw data for CNN learning were directly captured images for each species, Dataset 1 increases the number of images to improve the classification of fish species and Dataset 2 realizes images close to natural environment are constructed and used as training and test data. The classification performance of four CNNs are over 99.97% for dataset 1 and 99.5% for dataset 2, in particular, we confirm that the learned CNN using Data Set 2 has satisfactory performance for fish images similar to the natural environment. And among four CNNs, AlexNet achieves satisfactory performance, and this has also the shortest execution time and training time, we confirm that it is the most suitable structure to develop the system for the elimination of foreign fish species.

Image Filtering Method for an Effective Inverse Tone-mapping (효과적인 역 톤 매핑을 위한 필터링 기법)

  • Kang, Rahoon;Park, Bumjun;Jeong, Jechang
    • Journal of Broadcast Engineering
    • /
    • v.24 no.2
    • /
    • pp.217-226
    • /
    • 2019
  • In this paper, we propose a filtering method that can improve the results of inverse tone-mapping using guided image filter. Inverse tone-mapping techniques have been proposed that convert LDR images to HDR. Recently, many algorithms have been studied to convert single LDR images into HDR images using CNN. Among them, there exists an algorithm for restoring pixel information using CNN which learned to restore saturated region. The algorithm does not suppress the noise in the non-saturation region and cannot restore the detail in the saturated region. The proposed algorithm suppresses the noise in the non-saturated region and restores the detail of the saturated region using a WGIF in the input image, and then applies it to the CNN to improve the quality of the final image. The proposed algorithm shows a higher quantitative image quality index than the existing algorithms when the HDR quantitative image quality index was measured.

Armed person detection using Deep Learning (딥러닝 기반의 무기 소지자 탐지)

  • Kim, Geonuk;Lee, Minhun;Huh, Yoojin;Hwang, Gisu;Oh, Seoung-Jun
    • Journal of Broadcast Engineering
    • /
    • v.23 no.6
    • /
    • pp.780-789
    • /
    • 2018
  • Nowadays, gun crimes occur very frequently not only in public places but in alleyways around the world. In particular, it is essential to detect a person armed by a pistol to prevent those crimes since small guns, such as pistols, are often used for those crimes. Because conventional works for armed person detection have treated an armed person as a single object in an input image, their accuracy is very low. The reason for the low accuracy comes from the fact that the gunman is treated as a single object although the pistol is a relatively much smaller object than the person. To solve this problem, we propose a novel algorithm called APDA(Armed Person Detection Algorithm). APDA detects the armed person using in a post-processing the positions of both wrists and the pistol achieved by the CNN-based human body feature detection model and the pistol detection model, respectively. We show that APDA can provide both 46.3% better recall and 14.04% better precision than SSD-MobileNet.

Real-Time Fire Detection based on CNN and Grad-CAM (CNN과 Grad-CAM 기반의 실시간 화재 감지)

  • Kim, Young-Jin;Kim, Eun-Gyung
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.22 no.12
    • /
    • pp.1596-1603
    • /
    • 2018
  • Rapidly detecting and warning of fires is necessary for minimizing human injury and property damage. Generally, when fires occur, both the smoke and the flames are generated, so fire detection systems need to detect both the smoke and the flames. However, most fire detection systems only detect flames or smoke and have the disadvantage of slower processing speed due to additional preprocessing task. In this paper, we implemented a fire detection system which predicts the flames and the smoke at the same time by constructing a CNN model that supports multi-labeled classification. Also, the system can monitor the fire status in real time by using Grad-CAM which visualizes the position of classes based on the characteristics of CNN. Also, we tested our proposed system with 13 fire videos and got an average accuracy of 98.73% and 95.77% respectively for the flames and the smoke.

Sea Ice Type Classification with Optical Remote Sensing Data (광학영상에서의 해빙종류 분류 연구)

  • Chi, Junhwa;Kim, Hyun-cheol
    • Korean Journal of Remote Sensing
    • /
    • v.34 no.6_2
    • /
    • pp.1239-1249
    • /
    • 2018
  • Optical remote sensing sensors provide visually more familiar images than radar images. However, it is difficult to discriminate sea ice types in optical images using spectral information based machine learning algorithms. This study addresses two topics. First, we propose a semantic segmentation which is a part of the state-of-the-art deep learning algorithms to identify ice types by learning hierarchical and spatial features of sea ice. Second, we propose a new approach by combining of semi-supervised and active learning to obtain accurate and meaningful labels from unlabeled or unseen images to improve the performance of supervised classification for multiple images. Therefore, we successfully added new labels from unlabeled data to automatically update the semantic segmentation model. This should be noted that an operational system to generate ice type products from optical remote sensing data may be possible in the near future.

Bolt-Loosening Detection using Vision-Based Deep Learning Algorithm and Image Processing Method (영상기반 딥러닝 및 이미지 프로세싱 기법을 이용한 볼트풀림 손상 검출)

  • Lee, So-Young;Huynh, Thanh-Canh;Park, Jae-Hyung;Kim, Jeong-Tae
    • Journal of the Computational Structural Engineering Institute of Korea
    • /
    • v.32 no.4
    • /
    • pp.265-272
    • /
    • 2019
  • In this paper, a vision-based deep learning algorithm and image processing method are proposed to detect bolt-loosening in steel connections. To achieve this objective, the following approaches are implemented. First, a bolt-loosening detection method that includes regional convolutional neural network(RCNN)-based deep learning algorithm and Hough line transform(HLT)-based image processing algorithm are designed. The RCNN-based deep learning algorithm is developed to identify and crop bolts in a connection image. The HLT-based image processing algorithm is designed to estimate the bolt angles from the cropped bolt images. Then, the proposed vision-based method is evaluated for verifying bolt-loosening detection in a lab-scale girder connection. The accuracy of the RCNN-based bolt detector and HLT-based bolt angle estimator are examined with respect to various perspective distortions.

Rock Classification Prediction in Tunnel Excavation Using CNN (CNN 기법을 활용한 터널 암판정 예측기술 개발)

  • Kim, Hayoung;Cho, Laehun;Kim, Kyu-Sun
    • Journal of the Korean Geotechnical Society
    • /
    • v.35 no.9
    • /
    • pp.37-45
    • /
    • 2019
  • Quick identification of the condition of tunnel face and optimized determination of support patterns during tunnel excavation in underground construction projects help engineers prevent tunnel collapse and safely excavate tunnels. This study investigates a CNN technique for quick determination of rock quality classification depending on the condition of tunnel face, and presents the procedure for rock quality classification using a deep learning technique and the improved method for accurate prediction. The VGG16 model developed by tens of thousands prestudied images was used for deep learning, and 1,469 tunnel face images were used to classify the five types of rock quality condition. In this study, the prediction accuracy using this technique was up to 83.9%. It is expected that this technique can be used for an error-minimizing rock quality classification system not depending on experienced professionals in rock quality rating.

CycleGAN-based Object Detection under Night Environments (CycleGAN을 이용한 야간 상황 물체 검출 알고리즘)

  • Cho, Sangheum;Lee, Ryong;Na, Jaemin;Kim, Youngbin;Park, Minwoo;Lee, Sanghwan;Hwang, Wonjun
    • Journal of Korea Multimedia Society
    • /
    • v.22 no.1
    • /
    • pp.44-54
    • /
    • 2019
  • Recently, image-based object detection has made great progress with the introduction of Convolutional Neural Network (CNN). Many trials such as Region-based CNN, Fast R-CNN, and Faster R-CNN, have been proposed for achieving better performance in object detection. YOLO has showed the best performance under consideration of both accuracy and computational complexity. However, these data-driven detection methods including YOLO have the fundamental problem is that they can not guarantee the good performance without a large number of training database. In this paper, we propose a data sampling method using CycleGAN to solve this problem, which can convert styles while retaining the characteristics of a given input image. We will generate the insufficient data samples for training more robust object detection without efforts of collecting more database. We make extensive experimental results using the day-time and night-time road images and we validate the proposed method can improve the object detection accuracy of the night-time without training night-time object databases, because we converts the day-time training images into the synthesized night-time images and we train the detection model with the real day-time images and the synthesized night-time images.

Development and evaluation of AI-based algorithm models for analysis of learning trends in adult learners (성인 학습자의 학습 추이 분석을 위한 인공지능 기반 알고리즘 모델 개발 및 평가)

  • Jeong, Youngsik;Lee, Eunjoo;Do, Jaewoo
    • Journal of The Korean Association of Information Education
    • /
    • v.25 no.5
    • /
    • pp.813-824
    • /
    • 2021
  • To improve educational performance by analyzing the learning trends of adult learners of Open High Schools, various algorithm models using artificial intelligence were designed and performance was evaluated by applying them to real data. We analyzed Log data of 115 adult learners in the cyber education system of Open High Schools. Most adult learners of Open High Schools learned more than recommended learning time, but at the end of the semester, the actual learning time was significantly reduced compared to the recommended learning time. In the second half of learning, the participation rate of VODs, formation assessments, and learning activities also decreased. Therefore, in order to improve educational performance, learning time should be supported to continue in the second half. In the latter half, we developed an artificial intelligence algorithm models using Tensorflow to predict learning time by data they started taking the course. As a result, when using CNN(Convolutional Neural Network) model to predict single or multiple outputs, the mean-absolute-error is lowest compared to other models.

Indoor Scene Classification based on Color and Depth Images for Automated Reverberation Sound Editing (자동 잔향 편집을 위한 컬러 및 깊이 정보 기반 실내 장면 분류)

  • Jeong, Min-Heuk;Yu, Yong-Hyun;Park, Sung-Jun;Hwang, Seung-Jun;Baek, Joong-Hwan
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.24 no.3
    • /
    • pp.384-390
    • /
    • 2020
  • The reverberation effect on the sound when producing movies or VR contents is a very important factor in the realism and liveliness. The reverberation time depending the space is recommended in a standard called RT60(Reverberation Time 60 dB). In this paper, we propose a scene recognition technique for automatic reverberation editing. To this end, we devised a classification model that independently trains color images and predicted depth images in the same model. Indoor scene classification is limited only by training color information because of the similarity of internal structure. Deep learning based depth information extraction technology is used to use spatial depth information. Based on RT60, 10 scene classes were constructed and model training and evaluation were conducted. Finally, the proposed SCR + DNet (Scene Classification for Reverb + Depth Net) classifier achieves higher performance than conventional CNN classifiers with 92.4% accuracy.