• Title/Summary/Keyword: multi multi-layer neural networks

Search Result 225, Processing Time 0.023 seconds

Multi-Modal Wearable Sensor Integration for Daily Activity Pattern Analysis with Gated Multi-Modal Neural Networks (Gated Multi-Modal Neural Networks를 이용한 다중 웨어러블 센서 결합 방법 및 일상 행동 패턴 분석)

  • On, Kyoung-Woon;Kim, Eun-Sol;Zhang, Byoung-Tak
    • KIISE Transactions on Computing Practices
    • /
    • v.23 no.2
    • /
    • pp.104-109
    • /
    • 2017
  • We propose a new machine learning algorithm which analyzes daily activity patterns of users from multi-modal wearable sensor data. The proposed model learns and extracts activity patterns using input from wearable devices in real-time. Inspired by cue integration of human's property, we constructed gated multi-modal neural networks which integrate wearable sensor input data selectively by using gate modules. For the experiments, sensory data were collected by using multiple wearable devices in restaurant situations. As an experimental result, we first show that the proposed model performs well in terms of prediction accuracy. Then, the possibility to construct a knowledge schema automatically by analyzing the activation patterns in the middle layer of our proposed model is explained.

Improved Modeling of I-V Characteristic Based on Artificial Neural Network in Photovoltaic Systems (태양광 시스템의 인공신경망 기반 I-V 특성 모델링 향상)

  • Park, Jiwon;Lee, Jonghwan
    • Journal of the Semiconductor & Display Technology
    • /
    • v.21 no.3
    • /
    • pp.135-139
    • /
    • 2022
  • The current-voltage modeling plays an important role in characterizing photovoltaic systems. A solar cell has a nonlinear characteristic with various parameters influenced by the external environments such as the irradiance and the temperature. In order to accurately predict current-voltage characteristics at low irradiance, the artificial neural networks are applied to effectively quantify nonlinear behaviors. In this paper, a multi-layer perceptron scheme that can make accurate predictions is employed to learn complex formulas for large amounts of continuous data. The simulated results of artificial neural networks model show the accuracy improvement by using MATLAB/Simulink.

Multi-Layer Perceptron Based Ternary Tree Partitioning Decision Method for Versatile Video Coding (다목적 비디오 부/복호화를 위한 다층 퍼셉트론 기반 삼항 트리 분할 결정 방법)

  • Lee, Taesik;Jun, Dongsan
    • Journal of Korea Multimedia Society
    • /
    • v.25 no.6
    • /
    • pp.783-792
    • /
    • 2022
  • Versatile Video Coding (VVC) is the latest video coding standard, which had been developed by the Joint Video Experts Team (JVET) of ITU-T Video Coding Experts Group (VCEG) and ISO/IEC Moving Picture Experts Group (MPEG) in 2020. Although VVC can provide powerful coding performance, it requires tremendous computational complexity to determine the optimal block structures during the encoding process. In this paper, we propose a fast ternary tree decision method using two neural networks with 7 nodes as input vector based on the multi-layer perceptron structure, names STH-NN and STV-NN. As a training result of neural network, the STH-NN and STV-NN achieved accuracies of 85% and 91%, respectively. Experimental results show that the proposed method reduces the encoding complexity up to 25% with unnoticeable coding loss compared to the VVC test model (VTM).

Modular Neural Network Using Recurrent Neural Network (궤환 신경회로망을 사용한 모듈라 네트워크)

  • 최우경;김성주;서재용;전흥태
    • Proceedings of the IEEK Conference
    • /
    • 2003.07d
    • /
    • pp.1565-1568
    • /
    • 2003
  • In this paper, we propose modular network to solve difficult and complex problems that are seldom solved with multi-layer neural network. The structure of modular neural network in researched by Jacobs and Jordan is selected in this paper. Modular network consists of several expert networks and a gating network which is composed of single-layer neural network or multi-layer neural network. We propose modular network structure using recurrent neural network, since the state of the whole network at a particular time depends on an aggregate of previous states as well as on the current input. Finally, we show excellence of the proposed network compared with modular network.

  • PDF

Predicting the indirect tensile strength of self-compacting concrete using artificial neural networks

  • Mazloom, Moosa;Yoosefi, M.M.
    • Computers and Concrete
    • /
    • v.12 no.3
    • /
    • pp.285-301
    • /
    • 2013
  • This paper concentrates on the results of experimental work on tensile strength of self-compacting concrete (SCC) caused by flexure, which is called rupture modulus. The work focused on concrete mixes having water/binder ratios of 0.35 and 0.45, which contained constant total binder contents of 500 $kg/m^3$ and 400 $kg/m^3$, respectively. The concrete mixes had four different dosages of a superplasticizer based on polycarboxylic with and without silica fume. The percentage of silica fume that replaced cement in this research was 10%. Based upon the experimental results, the existing equations for anticipating the rupture modulus of SCC according to its compressive strength were not exact enough. Therefore, it is decided to use artificial neural networks (ANN) for anticipating the rupture modulus of SCC from its compressive strength and workability. The conclusion was that the multi layer perceptron (MLP) networks could predict the tensile strength in all conditions, but radial basis (RB) networks were not exact enough in some circumstances. On the other hand, RB networks were more users friendly and they converged to the final networks quicker.

A Study on Development of Embedded System for Speech Recognition using Multi-layer Recurrent Neural Prediction Models & HMM (다층회귀신경예측 모델 및 HMM 를 이용한 임베디드 음성인식 시스템 개발에 관한 연구)

  • Kim, Jung hoon;Jang, Won il;Kim, Young tak;Lee, Sang bae
    • Journal of the Korean Institute of Intelligent Systems
    • /
    • v.14 no.3
    • /
    • pp.273-278
    • /
    • 2004
  • In this paper, the recurrent neural networks (RNN) is applied to compensate for HMM recognition algorithm, which is commonly used as main recognizer. Among these recurrent neural networks, the multi-layer recurrent neural prediction model (MRNPM), which allows operating in real-time, is used to implement learning and recognition, and HMM and MRNPM are used to design a hybrid-type main recognizer. After testing the designed speech recognition algorithm with Korean number pronunciations (13 words), which are hardly distinct, for its speech-independent recognition ratio, about 5% improvement was obtained comparing with existing HMM recognizers. Based on this result, only optimal (recognition) codes were extracted in the actual DSP (TMS320C6711) environment, and the embedded speech recognition system was implemented. Similarly, the implementation result of the embedded system showed more improved recognition system implementation than existing solid HMM recognition systems.

Multi-layer Neural Network with Hybrid Learning Rules for Improved Robust Capability (Robustness를 형성시키기 위한 Hybrid 학습법칙을 갖는 다층구조 신경회로망)

  • 정동규;이수영
    • Journal of the Korean Institute of Telematics and Electronics B
    • /
    • v.31B no.8
    • /
    • pp.211-218
    • /
    • 1994
  • In this paper we develope a hybrid learning rule to improve the robustness of multi-layer Perceptions. In most neural networks the activation of a neuron is deternined by a nonlinear transformation of the weighted sum of inputs to the neurons. Investigating the behaviour of activations of hidden layer neurons a new learning algorithm is developed for improved robustness for multi-layer Perceptrons. Unlike other methods which reduce the network complexity by putting restrictions on synaptic weights our method based on error-backpropagation increases the complexity of the underlying proplem by imposing it saturation requirement on hidden layer neurons. We also found that the additional gradient-descent term for the requirement corresponds to the Hebbian rule and our algorithm incorporates the Hebbian learning rule into the error back-propagation rule. Computer simulation demonstrates fast learning convergence as well as improved robustness for classification and hetero-association of patterns.

  • PDF

Recurrent Based Modular Neural Network

  • Yon, Jung-Heum;Park, Woo-Kyung;Kim, Yong-Min;Jeon, Hong-Tae
    • Proceedings of the Korean Institute of Intelligent Systems Conference
    • /
    • 2003.09a
    • /
    • pp.694-697
    • /
    • 2003
  • In this paper, we propose modular network to solve difficult and complex problems that are seldom solved with Multi-Layer Neural Network(MLNN). The structure of Modular Neural Network(MNN) in researched by Jacobs and jordan is selected in this paper. Modular network consists of several Expert Networks(EN) and a Gating Network(CN) which is composed of single-layer neural network(SLNN) or multi-layer neural network. We propose modular network structure using Recurrent Neural Network(RNN), since the state of the whole network at a particular time depends on aggregate of previous states as well as on the current input. Finally, we show excellence of the proposed network compared with modular network.

  • PDF

Neural Network Training Using a GMDH Type Algorithm

  • Pandya, Abhijit S.;Gilbar, Thomas;Kim, Kwang-Baek
    • International Journal of Fuzzy Logic and Intelligent Systems
    • /
    • v.5 no.1
    • /
    • pp.52-58
    • /
    • 2005
  • We have developed a Group Method of Data Handling (GMDH) type algorithm for designing multi-layered neural networks. The algorithm is general enough that it will accept any number of inputs and any sized training set. Each neuron of the resulting network is a function of two of the inputs to the layer. The equation for each of the neurons is a quadratic polynomial. Several forms of the equation are tested for each neuron to make sure that only the best equation of two inputs is kept. All possible combinations of two inputs to each layer are also tested. By carefully testing each resulting neuron, we have developed an algorithm to keep only the best neurons at each level. The algorithm's goal is to create as accurate a network as possible while minimizing the size of the network. Software was developed to train and simulate networks using our algorithm. Several applications were modeled using our software, and the result was that our algorithm succeeded in developing small, accurate, multi-layer networks.

New Approach to Optimize the Size of Convolution Mask in Convolutional Neural Networks

  • Kwak, Young-Tae
    • Journal of the Korea Society of Computer and Information
    • /
    • v.21 no.1
    • /
    • pp.1-8
    • /
    • 2016
  • Convolutional neural network (CNN) consists of a few pairs of both convolution layer and subsampling layer. Thus it has more hidden layers than multi-layer perceptron. With the increased layers, the size of convolution mask ultimately determines the total number of weights in CNN because the mask is shared among input images. It also is an important learning factor which makes or breaks CNN's learning. Therefore, this paper proposes the best method to choose the convolution size and the number of layers for learning CNN successfully. Through our face recognition with vast learning examples, we found that the best size of convolution mask is 5 by 5 and 7 by 7, regardless of the number of layers. In addition, the CNN with two pairs of both convolution and subsampling layer is found to make the best performance as if the multi-layer perceptron having two hidden layers does.