• Title/Summary/Keyword: Deep learning algorithm

Search Result 1,146, Processing Time 0.024 seconds

Design and Implementation of a Face Authentication System (딥러닝 기반의 얼굴인증 시스템 설계 및 구현)

  • Lee, Seungik
    • Journal of Software Assessment and Valuation
    • /
    • v.16 no.2
    • /
    • pp.63-68
    • /
    • 2020
  • This paper proposes a face authentication system based on deep learning framework. The proposed system is consisted of face region detection and feature extraction using deep learning algorithm, and performed the face authentication using joint-bayesian matrix learning algorithm. The performance of proposed paper is evaluated by various face database , and the face image of one person consists of 2 images. The face authentication algorithm was performed by measuring similarity by applying 2048 dimension characteristic and combined Bayesian algorithm through Deep Neural network and calculating the same error rate that failed face certification. The result of proposed paper shows that the proposed system using deep learning and joint bayesian algorithms showed the equal error rate of 1.2%, and have a good performance compared to previous approach.

Optimization of Deep Learning Model Using Genetic Algorithm in PET-CT Image Alzheimer's Classification (PET-CT 영상 알츠하이머 분류에서 유전 알고리즘 이용한 심층학습 모델 최적화)

  • Lee, Sanghyeop;Kang, Do-Young;Song, Jongkwan;Park, Jangsik
    • Journal of Korea Multimedia Society
    • /
    • v.23 no.9
    • /
    • pp.1129-1138
    • /
    • 2020
  • The performance of convolutional deep learning networks is generally determined according to parameters of target dataset, structure of network, convolution kernel, activation function, and optimization algorithm. In this paper, a genetic algorithm is used to select the appropriate deep learning model and parameters for Alzheimer's classification and to compare the learning results with preliminary experiment. We compare and analyze the Alzheimer's disease classification performance of VGG-16, GoogLeNet, and ResNet to select an effective network for detecting AD and MCI. The simulation results show that the network structure is ResNet, the activation function is ReLU, the optimization algorithm is Adam, and the convolution kernel has a 3-dilated convolution filter for the accuracy of dementia medical images.

Deep-learning Sliding Window Based Object Detection and Tracking for Generating Trigger Signal of the LPR System (LPR 시스템 트리거 신호 생성을 위한 딥러닝 슬라이딩 윈도우 방식의 객체 탐지 및 추적)

  • Kim, Jinho
    • Journal of Korea Society of Digital Industry and Information Management
    • /
    • v.17 no.4
    • /
    • pp.85-94
    • /
    • 2021
  • The LPR system's trigger sensor makes problem occasionally due to the heave weight of vehicle or the obsolescence equipment. If we replace the hardware sensor to the deep-learning based software sensor in order to generate the trigger signal, LPR system maintenance would be a lot easier. In this paper we proposed the deep-learning sliding window based object detection and tracking algorithm for the LPR system's trigger signal generation. The gate passing vehicle's license plate recognition results are combined into the normal tracking algorithm to catch the position of the vehicle on the trigger line. The experimental results show that the deep learning sliding window based trigger signal generating performance was 100% for the gate passing vehicles including the 5.5% trigger signal position errors due to the minimum bounding box location errors in the vehicle detection process.

High-performance of Deep learning Colorization With Wavelet fusion (웨이블릿 퓨전에 의한 딥러닝 색상화의 성능 향상)

  • Kim, Young-Back;Choi, Hyun;Cho, Joong-Hwee
    • IEMEK Journal of Embedded Systems and Applications
    • /
    • v.13 no.6
    • /
    • pp.313-319
    • /
    • 2018
  • We propose a post-processing algorithm to improve the quality of the RGB image generated by deep learning based colorization from the gray-scale image of an infrared camera. Wavelet fusion is used to generate a new luminance component of the RGB image luminance component from the deep learning model and the luminance component of the infrared camera. PSNR is increased for all experimental images by applying the proposed algorithm to RGB images generated by two deep learning models of SegNet and DCGAN. For the SegNet model, the average PSNR is improved by 1.3906dB at level 1 of the Haar wavelet method. For the DCGAN model, PSNR is improved 0.0759dB on the average at level 5 of the Daubechies wavelet method. It is also confirmed that the edge components are emphasized by the post-processing and the visibility is improved.

A Basic Study on Estimation Method of Concrete Compressive Strength Based on Deep Learning Algorithm Considering Mixture Factor (배합 인자를 고려한 Deep Learning Algorithm을 이용한 콘크리트 압축강도 추정 기법에 관한 기초적 연구)

  • Lee, Seung-Jun;Kim, In-Soo;Lee, Han-Seung
    • Proceedings of the Korean Institute of Building Construction Conference
    • /
    • 2017.11a
    • /
    • pp.83-84
    • /
    • 2017
  • In the construction site, it is necessary to estimate the compressive strength of concrete in order to adjust the demolding time of the form, and establish and adjust the construction schedule. The compressive strength of concrete is determined by various influencing factors. However, the conventional method for estimating the compressive strength of concrete has been suggested by considering only 1 to 3 specific influential factors as variables. In this study, seven influential factors (W/B ratio, Water, Cement, Fly ash, Blast furnace slag, Curing temperature, and humidity) of papers opened for 10 years were collected at three conferences in order to know the various correlations among data and the tendency of data. The purpose of this paper is to estimate compressive strength more accurately by applying it to algorithm of the Deep learning.

  • PDF

Development of deep learning-based holographic ultrasound generation algorithm (딥러닝 기반 초음파 홀로그램 생성 알고리즘 개발)

  • Lee, Moon Hwan;Hwang, Jae Youn
    • The Journal of the Acoustical Society of Korea
    • /
    • v.40 no.2
    • /
    • pp.169-175
    • /
    • 2021
  • Recently, an ultrasound hologram and its applications have gained attention in the ultrasound research field. However, the determination technique of transmit signal phases, which generate a hologram, has not been significantly advanced from the previous algorithms which are time-consuming iterative methods. Thus, we applied the deep learning technique, which has been previously adopted to generate an optical hologram, to generate an ultrasound hologram. We further examined the Deep learning-based Holographic Ultrasound Generation algorithm (Deep-HUG). We implement the U-Net-based algorithm and examine its generalizability by training on a dataset, which consists of randomly distributed disks, and testing on the alphabets (A-Z). Furthermore, we compare the Deep-HUG with the previous algorithm in terms of computation time, accuracy, and uniformity. It was found that the accuracy and uniformity of the Deep-HUG are somewhat lower than those of the previous algorithm whereas the computation time is 190 times faster than that of the previous algorithm, demonstrating that Deep-HUG has potential as a useful technique to rapidly generate an ultrasound hologram for various applications.

Implementation and Verification of Deep Learning-based Automatic Object Tracking and Handy Motion Control Drone System (심층학습 기반의 자동 객체 추적 및 핸디 모션 제어 드론 시스템 구현 및 검증)

  • Kim, Youngsoo;Lee, Junbeom;Lee, Chanyoung;Jeon, Hyeri;Kim, Seungpil
    • IEMEK Journal of Embedded Systems and Applications
    • /
    • v.16 no.5
    • /
    • pp.163-169
    • /
    • 2021
  • In this paper, we implemented a deep learning-based automatic object tracking and handy motion control drone system and analyzed the performance of the proposed system. The drone system automatically detects and tracks targets by analyzing images obtained from the drone's camera using deep learning algorithms, consisting of the YOLO, the MobileNet, and the deepSORT. Such deep learning-based detection and tracking algorithms have both higher target detection accuracy and processing speed than the conventional color-based algorithm, the CAMShift. In addition, in order to facilitate the drone control by hand from the ground control station, we classified handy motions and generated flight control commands through motion recognition using the YOLO algorithm. It was confirmed that such a deep learning-based target tracking and drone handy motion control system stably track the target and can easily control the drone.

Enhanced Machine Learning Algorithms: Deep Learning, Reinforcement Learning, and Q-Learning

  • Park, Ji Su;Park, Jong Hyuk
    • Journal of Information Processing Systems
    • /
    • v.16 no.5
    • /
    • pp.1001-1007
    • /
    • 2020
  • In recent years, machine learning algorithms are continuously being used and expanded in various fields, such as facial recognition, signal processing, personal authentication, and stock prediction. In particular, various algorithms, such as deep learning, reinforcement learning, and Q-learning, are continuously being improved. Among these algorithms, the expansion of deep learning is rapidly changing. Nevertheless, machine learning algorithms have not yet been applied in several fields, such as personal authentication technology. This technology is an essential tool in the digital information era, walking recognition technology as promising biometrics, and technology for solving state-space problems. Therefore, algorithm technologies of deep learning, reinforcement learning, and Q-learning, which are typical machine learning algorithms in various fields, such as agricultural technology, personal authentication, wireless network, game, biometric recognition, and image recognition, are being improved and expanded in this paper.

A Study On Memory Optimization for Applying Deep Learning to PC (딥러닝을 PC에 적용하기 위한 메모리 최적화에 관한 연구)

  • Lee, Hee-Yeol;Lee, Seung-Ho
    • Journal of IKEEE
    • /
    • v.21 no.2
    • /
    • pp.136-141
    • /
    • 2017
  • In this paper, we propose an algorithm for memory optimization to apply deep learning to PC. The proposed algorithm minimizes the memory and computation processing time by reducing the amount of computation processing and data required in the conventional deep learning structure in a general PC. The algorithm proposed in this paper consists of three steps: a convolution layer configuration process using a random filter with discriminating power, a data reduction process using PCA, and a CNN structure creation using SVM. The learning process is not necessary in the convolution layer construction process using the discriminating random filter, thereby shortening the learning time of the overall deep learning. PCA reduces the amount of memory and computation throughput. The creation of the CNN structure using SVM maximizes the effect of reducing the amount of memory and computational throughput required. In order to evaluate the performance of the proposed algorithm, we experimented with Yale University's Extended Yale B face database. The results show that the algorithm proposed in this paper has a similar performance recognition rate compared with the existing CNN algorithm. And it was confirmed to be excellent. Based on the algorithm proposed in this paper, it is expected that a deep learning algorithm with many data and computation processes can be implemented in a general PC.

A Cascade-hybrid Recommendation Algorithm based on Collaborative Deep Learning Technique for Accuracy Improvement and Low Latency

  • Lee, Hyun-ho;Lee, Won-jin;Lee, Jae-dong
    • Journal of Korea Multimedia Society
    • /
    • v.23 no.1
    • /
    • pp.31-42
    • /
    • 2020
  • During the 4th Industrial Revolution, service platforms utilizing diverse contents are emerging, and research on recommended systems that can be customized to users to provide quality service is being conducted. hybrid recommendation systems that provide high accuracy recommendations are being researched in various domains, and various filtering techniques, machine learning, and deep learning are being applied to recommended systems. However, in a recommended service environment where data must be analyzed and processed real time, the accuracy of the recommendation is important, but the computational speed is also very important. Due to high level of model complexity, a hybrid recommendation system or a Deep Learning-based recommendation system takes a long time to calculate. In this paper, a Cascade-hybrid recommended algorithm is proposed that can reduce the computational time while maintaining the accuracy of the recommendation. The proposed algorithm was designed to reduce the complexity of the model and minimize the computational speed while processing sequentially, rather than using existing weights or using a hybrid recommendation technique handled in parallel. Therefore, through the algorithms in this paper, contents can be analyzed and recommended effectively and real time through services such as SNS environments or shared economy platforms.