• Title/Summary/Keyword: Single-Hidden-Layer Neural Network

Search Result 43, Processing Time 0.022 seconds

Neural Network Active Control of Structures with Earthquake Excitation

  • Cho Hyun Cheol;Fadali M. Sami;Saiidi M. Saiid;Lee Kwon Soon
    • International Journal of Control, Automation, and Systems
    • /
    • v.3 no.2
    • /
    • pp.202-210
    • /
    • 2005
  • This paper presents a new neural network control for nonlinear bridge systems with earthquake excitation. We design multi-layer neural network controllers with a single hidden layer. The selection of an optimal number of neurons in the hidden layer is an important design step for control performance. To select an optimal number of hidden neurons, we progressively add one hidden neuron and observe the change in a performance measure given by the weighted sum of the system error and the control force. The number of hidden neurons which minimizes the performance measure is selected for implementation. A neural network was trained for mitigating vibrations of bridge systems caused by El Centro earthquake. We applied the proposed control approach to a single-degree-of-freedom (SDOF) and a two-degree-of-freedom (TDOF) bridge system. We assessed the robustness of the control system using randomly generated earthquake excitations which were not used in training the neural network. Our results show that the neural network controller drastically mitigates the effect of the disturbance.

The Structure of Boundary Decision Using the Back Propagation Algorithms (역전파 알고리즘을 이용한 경계결정의 구성에 관한 연구)

  • Lee, Ji-Young
    • The Journal of Information Technology
    • /
    • v.8 no.1
    • /
    • pp.51-56
    • /
    • 2005
  • The Back propagation algorithm is a very effective supervised training method for multi-layer feed forward neural networks. This paper studies the decision boundary formation based on the Back propagation algorithm. The discriminating powers of several neural network topology are also investigated against five manually created data sets. It is found that neural networks with multiple hidden layer perform better than single hidden layer.

  • PDF

Multilayer Neural Network Using Delta Rule: Recognitron III (텔타규칙을 이용한 다단계 신경회로망 컴퓨터:Recognitron III)

  • 김춘석;박충규;이기한;황희영
    • The Transactions of the Korean Institute of Electrical Engineers
    • /
    • v.40 no.2
    • /
    • pp.224-233
    • /
    • 1991
  • The multilayer expanson of single layer NN (Neural Network) was needed to solve the linear seperability problem as shown by the classic example using the XOR function. The EBP (Error Back Propagation ) learning rule is often used in multilayer Neural Networks, but it is not without its faults: 1)D.Rimmelhart expanded the Delta Rule but there is a problem in obtaining Ca from the linear combination of the Weight matrix N between the hidden layer and the output layer and H, wich is the result of another linear combination between the input pattern and the Weight matrix M between the input layer and the hidden layer. 2) Even if using the difference between Ca and Da to adjust the values of the Weight matrix N between the hidden layer and the output layer may be valid is correct, but using the same value to adjust the Weight matrixd M between the input layer and the hidden layer is wrong. Recognitron III was proposed to solve these faults. According to simulation results, since Recognitron III does not learn the three layer NN itself, but divides it into several single layer NNs and learns these with learning patterns, the learning time is 32.5 to 72.2 time faster than EBP NN one. The number of patterns learned in a EBP NN with n input and output cells and n+1 hidden cells are 2**n, but n in Recognitron III of the same size. [5] In the case of pattern generalization, however, EBP NN is less than Recognitron III.

  • PDF

Fuzzy Supervised Learning Algorithm by using Self-generation (Self-generation을 이용한 퍼지 지도 학습 알고리즘)

  • 김광백
    • Journal of Korea Multimedia Society
    • /
    • v.6 no.7
    • /
    • pp.1312-1320
    • /
    • 2003
  • In this paper, we consider a multilayer neural network, with a single hidden layer. Error backpropagation learning method used widely in multilayer neural networks has a possibility of local minima due to the inadequate weights and the insufficient number of hidden nodes. So we propose a fuzzy supervised learning algorithm by using self-generation that self-generates hidden nodes by the compound fuzzy single layer perceptron and modified ART1. From the input layer to hidden layer, a modified ART1 is used to produce nodes. And winner take-all method is adopted to the connection weight adaptation, so that a stored pattern for some pattern gets updated. The proposed method has applied to the student identification card images. In simulation results, the proposed method reduces a possibility of local minima and improves learning speed and paralysis than the conventional error backpropagation learning algorithm.

  • PDF

Bayesian Analysis for Neural Network Models

  • Chung, Younshik;Jung, Jinhyouk;Kim, Chansoo
    • Communications for Statistical Applications and Methods
    • /
    • v.9 no.1
    • /
    • pp.155-166
    • /
    • 2002
  • Neural networks have been studied as a popular tool for classification and they are very flexible. Also, they are used for many applications of pattern classification and pattern recognition. This paper focuses on Bayesian approach to feed-forward neural networks with single hidden layer of units with logistic activation. In this model, we are interested in deciding the number of nodes of neural network model with p input units, one hidden layer with m hidden nodes and one output unit in Bayesian setup for fixed m. Here, we use the latent variable into the prior of the coefficient regression, and we introduce the 'sequential step' which is based on the idea of the data augmentation by Tanner and Wong(1787). The MCMC method(Gibbs sampler and Metropolish algorithm) can be used to overcome the complicated Bayesian computation. Finally, a proposed method is applied to a simulated data.

A Study on Handwritten Digit Recognition by Layer Combination of Multiple Neural Network (다중 신경망의 계층 결합에 의한 필기체 숫자 인식에 관한 연구)

  • 김두식;임길택;남윤석
    • Proceedings of the IEEK Conference
    • /
    • 1999.11a
    • /
    • pp.468-471
    • /
    • 1999
  • In this paper, we present a solution for combining multiple neural networks. Each neural network is trained with different features. And the neural networks are combined by four methods. The recognition rates by four combination methods are compared. The experimental results for handwritten digit recognition shows that the combination at hidden layers by single layer neural network is superior to any other methods. The reasons of the results are explained.

  • PDF

Classification Performance Improvement of Steam Generator Tube Defects in Nuclear Power Plant Using Bagging Method (Bagging 방법을 이용한 원전SG 세관 결함패턴 분류성능 향상기법)

  • Lee, Jun-Po;Jo, Nam-Hoon
    • The Transactions of The Korean Institute of Electrical Engineers
    • /
    • v.58 no.12
    • /
    • pp.2532-2537
    • /
    • 2009
  • For defect characterization in steam generator tubes in nuclear power plant, artificial neural network has been extensively used to classify defect types. In this paper, we study the effectiveness of Bagging for improving the performance of neural network for the classification of tube defects. Bagging is a method that combines outputs of many neural networks that were trained separately with different training data set. By varying the number of neurons in the hidden layer, we carry out computer simulations in order to compare the classification performance of bagging neural network and single neural network. From the experiments, we found that the performance of bagging neural network is superior to the average performance of single neural network in most cases.

A Study on an Artificial Neural Network Design using Evolutionary Programming (진화 프로그래밍 기법을 이용한 신경망의 자동설계에 관한 연구)

  • 강신준;고택범;우천희;이덕규;우광방
    • Journal of Institute of Control, Robotics and Systems
    • /
    • v.5 no.3
    • /
    • pp.281-287
    • /
    • 1999
  • In this paper, a design method based on evolutionary programming for feedforward neural networks which have a single hidden layer is presented. By using an evolutionary programming, the network parameters such as the network structure, weight, slope of sigmoid functions and bias of nodes can be acquired simultaneously. To check the effectiveness of the suggested method, two numerical examples are examined. The performance of the identified network is demonstrated.

  • PDF

Neural Network Image Reconstruction for Magnetic Particle Imaging

  • Chae, Byung Gyu
    • ETRI Journal
    • /
    • v.39 no.6
    • /
    • pp.841-850
    • /
    • 2017
  • We investigate neural network image reconstruction for magnetic particle imaging. The network performance strongly depends on the convolution effects of the spectrum input data. The larger convolution effect appearing at a relatively smaller nanoparticle size obstructs the network training. The trained single-layer network reveals the weighting matrix consisting of a basis vector in the form of Chebyshev polynomials of the second kind. The weighting matrix corresponds to an inverse system matrix, where an incoherency of basis vectors due to low convolution effects, as well as a nonlinear activation function, plays a key role in retrieving the matrix elements. Test images are well reconstructed through trained networks having an inverse kernel matrix. We also confirm that a multi-layer network with one hidden layer improves the performance. Based on the results, a neural network architecture overcoming the low incoherence of the inverse kernel through the classification property is expected to become a better tool for image reconstruction.

Prediction of Arc Welding Quality through Artificial Neural Network (신경망 알고리즘을 이용한 아크 용접부 품질 예측)

  • Cho, Jungho
    • Journal of Welding and Joining
    • /
    • v.31 no.3
    • /
    • pp.44-48
    • /
    • 2013
  • Artificial neural network (ANN) model is applied to predict arc welding process window for automotive steel plate. Target weldment was various automotive steel plate combination with lap fillet joint. The accuracy of prediction was evaluated through comparison experimental result to ANN simulation. The effect of ANN variables on the accuracy is investigated such as number of hidden layers, perceptrons and transfer function type. A static back propagation model is established and tested. The result shows comparatively accurate predictability of the suggested ANN model. However, it restricts to use nonlinear transfer function instead of linear type and suggests only one single hidden layer rather than multiple ones to get better accuracy. In addition to this, obvious fact is affirmed again that the more perceptrons guarantee the better accuracy under the precondition that there are enough experimental database to train the neural network.