• Title/Summary/Keyword: Multiple Inputs Deep Neural Networks

Search Result 4, Processing Time 0.017 seconds

Multiple Inputs Deep Neural Networks for Bone Age Estimation Using Whole-Body Bone Scintigraphy

  • Nguyen, Phap Do Cong;Baek, Eu-Tteum;Yang, Hyung-Jeong;Kim, Soo-Hyung;Kang, Sae-Ryung;Min, Jung-Joon
    • Journal of Korea Multimedia Society
    • /
    • v.22 no.12
    • /
    • pp.1376-1384
    • /
    • 2019
  • The cosmetic and behavioral aspects of aging have become increasingly evident over the years. Physical aging in people can easily be observed on their face, posture, voice, and gait. In contrast, bone aging only becomes apparent once significant bone degeneration manifests through degenerative bone diseases. Therefore, a more accurate and timely assessment of bone aging is needed so that the determinants and its mechanisms can be more effectively identified and ultimately optimized. This study proposed a deep learning approach to assess the bone age of an adult using whole-body bone scintigraphy. The proposed approach uses multiple inputs deep neural network architectures using a loss function, called mean-variance loss. The data set was collected from Chonnam National University Hwasun Hospital. The experiment results show the effectiveness of the proposed method with a mean absolute error of 3.40 years.

Faults detection and identification for gas turbine using DNN and LLM

  • Oliaee, Seyyed Mohammad Emad;Teshnehlab, Mohammad;Shoorehdeli, Mahdi Aliyari
    • Smart Structures and Systems
    • /
    • v.23 no.4
    • /
    • pp.393-403
    • /
    • 2019
  • Applying more features gives us better accuracy in modeling; however, increasing the inputs causes the curse of dimensions. In this paper, a new structure has been proposed for fault detecting and identifying (FDI) of high-dimensional systems. This structure consist of two structure. The first part includes Auto-Encoders (AE) as Deep Neural Networks (DNNs) to produce feature engineering process and summarize the features. The second part consists of the Local Model Networks (LMNs) with LOcally LInear MOdel Tree (LOLIMOT) algorithm to model outputs (multiple models). The fault detection is based on these multiple models. Hence the residuals generated by comparing the system output and multiple models have been used to alarm the faults. To show the effectiveness of the proposed structure, it is tested on single-shaft industrial gas turbine prototype model. Finally, a brief comparison between the simulated results and several related works is presented and the well performance of the proposed structure has been illustrated.

Deep Learning Architectures and Applications (딥러닝의 모형과 응용사례)

  • Ahn, SungMahn
    • Journal of Intelligence and Information Systems
    • /
    • v.22 no.2
    • /
    • pp.127-142
    • /
    • 2016
  • Deep learning model is a kind of neural networks that allows multiple hidden layers. There are various deep learning architectures such as convolutional neural networks, deep belief networks and recurrent neural networks. Those have been applied to fields like computer vision, automatic speech recognition, natural language processing, audio recognition and bioinformatics where they have been shown to produce state-of-the-art results on various tasks. Among those architectures, convolutional neural networks and recurrent neural networks are classified as the supervised learning model. And in recent years, those supervised learning models have gained more popularity than unsupervised learning models such as deep belief networks, because supervised learning models have shown fashionable applications in such fields mentioned above. Deep learning models can be trained with backpropagation algorithm. Backpropagation is an abbreviation for "backward propagation of errors" and a common method of training artificial neural networks used in conjunction with an optimization method such as gradient descent. The method calculates the gradient of an error function with respect to all the weights in the network. The gradient is fed to the optimization method which in turn uses it to update the weights, in an attempt to minimize the error function. Convolutional neural networks use a special architecture which is particularly well-adapted to classify images. Using this architecture makes convolutional networks fast to train. This, in turn, helps us train deep, muti-layer networks, which are very good at classifying images. These days, deep convolutional networks are used in most neural networks for image recognition. Convolutional neural networks use three basic ideas: local receptive fields, shared weights, and pooling. By local receptive fields, we mean that each neuron in the first(or any) hidden layer will be connected to a small region of the input(or previous layer's) neurons. Shared weights mean that we're going to use the same weights and bias for each of the local receptive field. This means that all the neurons in the hidden layer detect exactly the same feature, just at different locations in the input image. In addition to the convolutional layers just described, convolutional neural networks also contain pooling layers. Pooling layers are usually used immediately after convolutional layers. What the pooling layers do is to simplify the information in the output from the convolutional layer. Recent convolutional network architectures have 10 to 20 hidden layers and billions of connections between units. Training deep learning networks has taken weeks several years ago, but thanks to progress in GPU and algorithm enhancement, training time has reduced to several hours. Neural networks with time-varying behavior are known as recurrent neural networks or RNNs. A recurrent neural network is a class of artificial neural network where connections between units form a directed cycle. This creates an internal state of the network which allows it to exhibit dynamic temporal behavior. Unlike feedforward neural networks, RNNs can use their internal memory to process arbitrary sequences of inputs. Early RNN models turned out to be very difficult to train, harder even than deep feedforward networks. The reason is the unstable gradient problem such as vanishing gradient and exploding gradient. The gradient can get smaller and smaller as it is propagated back through layers. This makes learning in early layers extremely slow. The problem actually gets worse in RNNs, since gradients aren't just propagated backward through layers, they're propagated backward through time. If the network runs for a long time, that can make the gradient extremely unstable and hard to learn from. It has been possible to incorporate an idea known as long short-term memory units (LSTMs) into RNNs. LSTMs make it much easier to get good results when training RNNs, and many recent papers make use of LSTMs or related ideas.

Few-Shot Image Synthesis using Noise-Based Deep Conditional Generative Adversarial Nets

  • Msiska, Finlyson Mwadambo;Hassan, Ammar Ul;Choi, Jaeyoung;Yoo, Jaewon
    • Smart Media Journal
    • /
    • v.10 no.1
    • /
    • pp.79-87
    • /
    • 2021
  • In recent years research on automatic font generation with machine learning mainly focus on using transformation-based methods, in comparison, generative model-based methods of font generation have received less attention. Transformation-based methods learn a mapping of the transformations from an existing input to a target. This makes them ambiguous because in some cases a single input reference may correspond to multiple possible outputs. In this work, we focus on font generation using the generative model-based methods which learn the buildup of the characters from noise-to-image. We propose a novel way to train a conditional generative deep neural model so that we can achieve font style control on the generated font images. Our research demonstrates how to generate new font images conditioned on both character class labels and character style labels when using the generative model-based methods. We achieve this by introducing a modified generator network which is given inputs noise, character class, and style, which help us to calculate losses separately for the character class labels and character style labels. We show that adding the character style vector on top of the character class vector separately gives the model rich information about the font and enables us to explicitly specify not only the character class but also the character style that we want the model to generate.