• Title/Summary/Keyword: Deep neural networks

Search Result 854, Processing Time 0.028 seconds

Learning Less Random to Learn Better in Deep Reinforcement Learning with Noisy Parameters

  • Kim, Chayoung
    • Journal of Advanced Information Technology and Convergence
    • /
    • v.9 no.1
    • /
    • pp.127-134
    • /
    • 2019
  • In terms of deep Reinforcement Learning (RL), exploration can be worked stochastically in the action of a state space. On the other hands, exploitation can be done the proportion of well generalization behaviors. The balance of exploration and exploitation is extremely important for better results. The randomly selected action with ε-greedy for exploration has been regarded as a de facto method. There is an alternative method to add noise parameters into a neural network for richer exploration. However, it is not easy to predict or detect over-fitting with the stochastically exploration in the perturbed neural network. Moreover, the well-trained agents in RL do not necessarily prevent or detect over-fitting in the neural network. Therefore, we suggest a novel design of a deep RL by the balance of the exploration with drop-out to reduce over-fitting in the perturbed neural networks.

A USB classification system using deep neural networks (인공신경망을 이용한 USB 인식 시스템)

  • Woo, Sae-Hyeong;Park, Jisu;Eun, Seongbae;Cha, Shin
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2022.05a
    • /
    • pp.535-538
    • /
    • 2022
  • For Plug & Play of IoT devices, we develop a module that recognizes the type of USB, which is a typical wired interface of IoT devices, through image recognition. In order to drive an IoT device, a driver for communication and device hardware is required. The wired interface for connecting to the IoT device is recognized by using the image obtained through the camera of smartphone shooting to recognize the corresponding communication interface. For USB, which is a most popular wired interface, types of USB are classified through artificial neural network-based machine learning. In order to secure sufficient data set of artificial neural networks, USB images are collected through the Internet, and additional image data sets are secured through image processing. In addition to the convolution neural networks, recognizers are implemented with various deep artificial neural networks, and their performance is compared and evaluated.

  • PDF

The Method of Abandoned Object Recognition based on Neural Networks (신경망 기반의 유기된 물체 인식 방법)

  • Ryu, Dong-Gyun;Lee, Jae-Heung
    • Journal of IKEEE
    • /
    • v.22 no.4
    • /
    • pp.1131-1139
    • /
    • 2018
  • This paper proposes a method of recognition abandoned objects using convolutional neural networks. The method first detects an area for an abandoned object in image and, if there is a detected area, applies convolutional neural networks to that area to recognize which object is represented. Experiments were conducted through an application system that detects illegal trash dumping. The experiments result showed the area of abandoned object was detected efficiently. The detected areas enter the input of convolutional neural networks and are classified into whether it is a trash or not. To do this, I trained convolutional neural networks with my own trash dataset and open database. As a training result, I achieved high accuracy for the test set not included in the training set.

EPS Gesture Signal Recognition using Deep Learning Model (심층 학습 모델을 이용한 EPS 동작 신호의 인식)

  • Lee, Yu ra;Kim, Soo Hyung;Kim, Young Chul;Na, In Seop
    • Smart Media Journal
    • /
    • v.5 no.3
    • /
    • pp.35-41
    • /
    • 2016
  • In this paper, we propose hand-gesture signal recognition based on EPS(Electronic Potential Sensor) using Deep learning model. Extracted signals which from Electronic field based sensor, EPS have much of the noise, so it must remove in pre-processing. After the noise are removed with filter using frequency feature, the signals are reconstructed with dimensional transformation to overcome limit which have just one-dimension feature with voltage value for using convolution operation. Then, the reconstructed signal data is finally classified and recognized using multiple learning layers model based on deep learning. Since the statistical model based on probability is sensitive to initial parameters, the result can change after training in modeling phase. Deep learning model can overcome this problem because of several layers in training phase. In experiment, we used two different deep learning structures, Convolutional neural networks and Recurrent Neural Network and compared with statistical model algorithm with four kinds of gestures. The recognition result of method using convolutional neural network is better than other algorithms in EPS gesture signal recognition.

A Parallel Deep Convolutional Neural Network for Alzheimer's disease classification on PET/CT brain images

  • Baydargil, Husnu Baris;Park, Jangsik;Kang, Do-Young;Kang, Hyun;Cho, Kook
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.14 no.9
    • /
    • pp.3583-3597
    • /
    • 2020
  • In this paper, a parallel deep learning model using a convolutional neural network and a dilated convolutional neural network is proposed to classify Alzheimer's disease with high accuracy in PET/CT images. The developed model consists of two pipelines, a conventional CNN pipeline, and a dilated convolution pipeline. An input image is sent through both pipelines, and at the end of both pipelines, extracted features are concatenated and used for classifying Alzheimer's disease. Complimentary abilities of both networks provide better overall accuracy than single conventional CNNs in the dataset. Moreover, instead of performing binary classification, the proposed model performs three-class classification being Alzheimer's disease, mild cognitive impairment, and normal control. Using the data received from Dong-a University, the model performs classification detecting Alzheimer's disease with an accuracy of up to 95.51%.

Prediction of golden time for recovering SISs using deep fuzzy neural networks with rule-dropout

  • Jo, Hye Seon;Koo, Young Do;Park, Ji Hun;Oh, Sang Won;Kim, Chang-Hwoi;Na, Man Gyun
    • Nuclear Engineering and Technology
    • /
    • v.53 no.12
    • /
    • pp.4014-4021
    • /
    • 2021
  • If safety injection systems (SISs) do not work in the event of a loss-of-coolant accident (LOCA), the accident can progress to a severe accident in which the reactor core is exposed and the reactor vessel fails. Therefore, it is considered that a technology that provides recoverable maximum time for SIS actuation is necessary to prevent this progression. In this study, the corresponding time was defined as the golden time. To achieve the objective of accurately predicting the golden time, the prediction was performed using the deep fuzzy neural network (DFNN) with rule-dropout. The DFNN with rule-dropout has an architecture in which many of the fuzzy neural networks (FNNs) are connected and is a method in which the fuzzy rule numbers, which are directly related to the number of nodes in the FNN that affect inference performance, are properly adjusted by a genetic algorithm. The golden time prediction performance of the DFNN model with rule-dropout was better than that of the support vector regression model. By using the prediction result through the proposed DFNN with rule-dropout, it is expected to prevent the aggravation of the accidents by providing the maximum remaining time for SIS recovery, which failed in the LOCA situation.

Vehicle Image Recognition Using Deep Convolution Neural Network and Compressed Dictionary Learning

  • Zhou, Yanyan
    • Journal of Information Processing Systems
    • /
    • v.17 no.2
    • /
    • pp.411-425
    • /
    • 2021
  • In this paper, a vehicle recognition algorithm based on deep convolutional neural network and compression dictionary is proposed. Firstly, the network structure of fine vehicle recognition based on convolutional neural network is introduced. Then, a vehicle recognition system based on multi-scale pyramid convolutional neural network is constructed. The contribution of different networks to the recognition results is adjusted by the adaptive fusion method that adjusts the network according to the recognition accuracy of a single network. The proportion of output in the network output of the entire multiscale network. Then, the compressed dictionary learning and the data dimension reduction are carried out using the effective block structure method combined with very sparse random projection matrix, which solves the computational complexity caused by high-dimensional features and shortens the dictionary learning time. Finally, the sparse representation classification method is used to realize vehicle type recognition. The experimental results show that the detection effect of the proposed algorithm is stable in sunny, cloudy and rainy weather, and it has strong adaptability to typical application scenarios such as occlusion and blurring, with an average recognition rate of more than 95%.

Applying Deep Reinforcement Learning to Improve Throughput and Reduce Collision Rate in IEEE 802.11 Networks

  • Ke, Chih-Heng;Astuti, Lia
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.16 no.1
    • /
    • pp.334-349
    • /
    • 2022
  • The effectiveness of Wi-Fi networks is greatly influenced by the optimization of contention window (CW) parameters. Unfortunately, the conventional approach employed by IEEE 802.11 wireless networks is not scalable enough to sustain consistent performance for the increasing number of stations. Yet, it is still the default when accessing channels for single-users of 802.11 transmissions. Recently, there has been a spike in attempts to enhance network performance using a machine learning (ML) technique known as reinforcement learning (RL). Its advantage is interacting with the surrounding environment and making decisions based on its own experience. Deep RL (DRL) uses deep neural networks (DNN) to deal with more complex environments (such as continuous state spaces or actions spaces) and to get optimum rewards. As a result, we present a new approach of CW control mechanism, which is termed as contention window threshold (CWThreshold). It uses the DRL principle to define the threshold value and learn optimal settings under various network scenarios. We demonstrate our proposed method, known as a smart exponential-threshold-linear backoff algorithm with a deep Q-learning network (SETL-DQN). The simulation results show that our proposed SETL-DQN algorithm can effectively improve the throughput and reduce the collision rates.

Multiple Inputs Deep Neural Networks for Bone Age Estimation Using Whole-Body Bone Scintigraphy

  • Nguyen, Phap Do Cong;Baek, Eu-Tteum;Yang, Hyung-Jeong;Kim, Soo-Hyung;Kang, Sae-Ryung;Min, Jung-Joon
    • Journal of Korea Multimedia Society
    • /
    • v.22 no.12
    • /
    • pp.1376-1384
    • /
    • 2019
  • The cosmetic and behavioral aspects of aging have become increasingly evident over the years. Physical aging in people can easily be observed on their face, posture, voice, and gait. In contrast, bone aging only becomes apparent once significant bone degeneration manifests through degenerative bone diseases. Therefore, a more accurate and timely assessment of bone aging is needed so that the determinants and its mechanisms can be more effectively identified and ultimately optimized. This study proposed a deep learning approach to assess the bone age of an adult using whole-body bone scintigraphy. The proposed approach uses multiple inputs deep neural network architectures using a loss function, called mean-variance loss. The data set was collected from Chonnam National University Hwasun Hospital. The experiment results show the effectiveness of the proposed method with a mean absolute error of 3.40 years.

Video Expression Recognition Method Based on Spatiotemporal Recurrent Neural Network and Feature Fusion

  • Zhou, Xuan
    • Journal of Information Processing Systems
    • /
    • v.17 no.2
    • /
    • pp.337-351
    • /
    • 2021
  • Automatically recognizing facial expressions in video sequences is a challenging task because there is little direct correlation between facial features and subjective emotions in video. To overcome the problem, a video facial expression recognition method using spatiotemporal recurrent neural network and feature fusion is proposed. Firstly, the video is preprocessed. Then, the double-layer cascade structure is used to detect a face in a video image. In addition, two deep convolutional neural networks are used to extract the time-domain and airspace facial features in the video. The spatial convolutional neural network is used to extract the spatial information features from each frame of the static expression images in the video. The temporal convolutional neural network is used to extract the dynamic information features from the optical flow information from multiple frames of expression images in the video. A multiplication fusion is performed with the spatiotemporal features learned by the two deep convolutional neural networks. Finally, the fused features are input to the support vector machine to realize the facial expression classification task. The experimental results on cNTERFACE, RML, and AFEW6.0 datasets show that the recognition rates obtained by the proposed method are as high as 88.67%, 70.32%, and 63.84%, respectively. Comparative experiments show that the proposed method obtains higher recognition accuracy than other recently reported methods.