• Title/Summary/Keyword: Error Backpropagation Learning Algorithm

Search Result 60, Processing Time 0.029 seconds

Blending Precess Optimization using Fuzzy Set Theory an Neural Networks (퍼지 및 신경망을 이용한 Blending Process의 최적화)

  • 황인창;김정남;주관정
    • Proceedings of the Korean Society of Precision Engineering Conference
    • /
    • 1993.10a
    • /
    • pp.488-492
    • /
    • 1993
  • This paper proposes a new approach to the optimization method of a blending process with neural network. The method is based on the error backpropagation learning algorithm for neural network. Since the neural network can model an arbitrary nonlinear mapping, it is used as a system solver. A fuzzy membership function is used in parallel with the neural network to minimize the difference between measurement value and input value of neural network. As a result, we can guarantee the reliability and stability of blending process by the help of neural network and fuzzy membership function.

  • PDF

Adaptive Control of Nonlinear Systems through Improvement of Learning Speed of Neural Networks and Compensation of Control Inputs (신경망의 학습속도 개선 및 제어입력 보상을 통한 비선형 시스템의 적응제어)

  • 배병우;전기준
    • The Transactions of the Korean Institute of Electrical Engineers
    • /
    • v.43 no.6
    • /
    • pp.991-1000
    • /
    • 1994
  • To control nonlinear systems adaptively, we improve learning speed of neural networks and present a novel control algorithm characterized by compensation of control inputs. In an error-backpropagation algorithm for tranining multilayer neural networks(MLNN's) the effect of the slope of activation functions on learning performance is investigated and the learning speed of neural networks is improved by auto-adjusting the slope of activation functions. The control system is composed of two MLNN's, one for control and the other for identification, with the weights initialized by off-line training. The control algoritm is modified by a control strategy which compensates the control error induced by the indentification error. Computer simulations show that the proposed control algorithm is efficient in controlling a nonlinear system with abruptly changing parameters.

Identification of suspension systems using error self recurrent neural network and development of sliding mode controller (오차 자기 순환 신경회로망을 이용한 현가시스템 인식과 슬라이딩 모드 제어기 개발)

  • 송광현;이창구;김성중
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 1997.10a
    • /
    • pp.625-628
    • /
    • 1997
  • In this paper the new neural network and sliding mode suspension controller is proposed. That neural network is error self-recurrent neural network. For fast on-line learning, this paper use recursive least squares method. A new neural networks converges considerably faster than the backpropagation algorithm and has advantages of being less affected by the poor initial weights and learning rate. The controller for suspension systems is designed according to sliding mode technique based on new proposed neural network.

  • PDF

ART1-based Fuzzy Supervised Learning Algorithm (ART-1 기반 퍼지 지도 학습 알고리즘)

  • Kim Kwang-Baek;Cho Jae-Hyun
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.9 no.4
    • /
    • pp.883-889
    • /
    • 2005
  • Error backpropagation algorithm of multilayer perceptron may result in local-minima because of the insufficient nodes in the hidden layer, inadequate momentum set-up, and initial weights. In this paper, we proposed the ART-1 based fuzzy supervised learning algorithm which is composed of ART-1 and fuzzy single layer supervised learning algorithm. The Proposed fuzzy supervised learning algorithm using self-generation method applied not only ART-1 to creation of nodes from the input layer to the hidden layer, but also the winer-take-all method, modifying stored patterns according to specific patterns. to adjustment of weights. We have applied the proposed learning method to the problem of recognizing a resident registration number in resident cards. Our experimental result showed that the possibility of local-minima was decreased and the teaming speed and the paralysis were improved more than the conventional error backpropagation algorithm.

Batch-mode Learning in Neural Networks (신경회로망에서 일괄 학습)

  • 김명찬;최종호
    • Journal of the Korean Institute of Telematics and Electronics B
    • /
    • v.32B no.3
    • /
    • pp.503-511
    • /
    • 1995
  • A batch-mode algorithm is proposed to increase the speed of learning in the error backpropagation algorithm with variable learning rate and variable momentum parameters in classification problems. The objective function is normalized with respect to the number of patterns and output nodes. Also the gradient of the objective function is normalized in updating the connection weights to increase the effect of its backpropagated error. The learning rate and momentum parameters are determined from a function of the gradient norm and the number of weights. The learning rate depends on the square rott of the gradient norm while the momentum parameters depend on the gradient norm. In the two typical classification problems, simulation results demonstrate the effectiveness of the proposed algorithm.

  • PDF

An Improvement of the MLP Based Speaker Verification System through Improving the learning Speed and Reducing the Learning Data (학습속도 개선과 학습데이터 축소를 통한 MLP 기반 화자증명 시스템의 등록속도 향상방법)

  • Lee, Baek-Yeong;Lee, Tae-Seung;Hwang, Byeong-Won
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.39 no.3
    • /
    • pp.88-98
    • /
    • 2002
  • The multilayer perceptron (MLP) has several advantages against other pattern recognition methods, and is expected to be used as the learning and recognizing speakers of speaker verification system. But because of the low learning speed of the error backpropagation (EBP) algorithm that is used for the MLP learning, the MLP learning requires considerable time. Because the speaker verification system must provide verification services just after a speaker's enrollment, it is required to solve the problem. So, this paper tries to make short of time required to enroll speakers with the MLP based speaker verification system, using the method of improving the EBP learning speed and the method of reducing background speakers which adopts the cohort speakers method from the existing speaker verification.

Accelerating Levenberg-Marquardt Algorithm using Variable Damping Parameter (가변 감쇠 파라미터를 이용한 Levenberg-Marquardt 알고리즘의 학습 속도 향상)

  • Kwak, Young-Tae
    • Journal of the Korea Society of Computer and Information
    • /
    • v.15 no.4
    • /
    • pp.57-63
    • /
    • 2010
  • The damping parameter of Levenberg-Marquardt algorithm switches between error backpropagation and Gauss-Newton learning and affects learning speed. Fixing the damping parameter induces some oscillation of error and decreases learning speed. Therefore, we propose the way of a variable damping parameter with referring to the alternation of error. The proposed method makes the damping parameter increase if error rate is large and makes it decrease if error rate is small. This method so plays the role of momentum that it can improve learning speed. We tested both iris recognition and wine recognition for this paper. We found out that this method improved learning speed in 67% cases on iris recognition and in 78% cases on wine recognition. It was also showed that the oscillation of error by the proposed way was less than those of other algorithms.

Multi-layer Neural Network with Hybrid Learning Rules for Improved Robust Capability (Robustness를 형성시키기 위한 Hybrid 학습법칙을 갖는 다층구조 신경회로망)

  • 정동규;이수영
    • Journal of the Korean Institute of Telematics and Electronics B
    • /
    • v.31B no.8
    • /
    • pp.211-218
    • /
    • 1994
  • In this paper we develope a hybrid learning rule to improve the robustness of multi-layer Perceptions. In most neural networks the activation of a neuron is deternined by a nonlinear transformation of the weighted sum of inputs to the neurons. Investigating the behaviour of activations of hidden layer neurons a new learning algorithm is developed for improved robustness for multi-layer Perceptrons. Unlike other methods which reduce the network complexity by putting restrictions on synaptic weights our method based on error-backpropagation increases the complexity of the underlying proplem by imposing it saturation requirement on hidden layer neurons. We also found that the additional gradient-descent term for the requirement corresponds to the Hebbian rule and our algorithm incorporates the Hebbian learning rule into the error back-propagation rule. Computer simulation demonstrates fast learning convergence as well as improved robustness for classification and hetero-association of patterns.

  • PDF

Selective Attentive Learning for Fast Speaker Adaptation in Multilayer Perceptron (다층 퍼셉트론에서의 빠른 화자 적응을 위한 선택적 주의 학습)

  • 김인철;진성일
    • The Journal of the Acoustical Society of Korea
    • /
    • v.20 no.4
    • /
    • pp.48-53
    • /
    • 2001
  • In this paper, selectively attentive learning method has been proposed to improve the learning speed of multilayer Perceptron based on the error backpropagation algorithm. Three attention criterions are introduced to effectively determine which set of input patterns is or which portion of network is attended to for effective learning. Such criterions are based on the mean square error function of the output layer and class-selective relevance of the hidden nodes. The acceleration of learning time is achieved by lowering the computational cost per iteration. Effectiveness of the proposed method is demonstrated in a speaker adaptation task of isolated word recognition system. The experimental results show that the proposed selective attention technique can reduce the learning time more than 60% in an average sense.

  • PDF

An Adaptive Learning Rate with Limited Error Signals for Training of Multilayer Perceptrons

  • Oh, Sang-Hoon;Lee, Soo-Young
    • ETRI Journal
    • /
    • v.22 no.3
    • /
    • pp.10-18
    • /
    • 2000
  • Although an n-th order cross-entropy (nCE) error function resolves the incorrect saturation problem of conventional error backpropagation (EBP) algorithm, performance of multilayer perceptrons (MLPs) trained using the nCE function depends heavily on the order of nCE. In this paper, we propose an adaptive learning rate to markedly reduce the sensitivity of MLP performance to the order of nCE. Additionally, we propose to limit error signal values at out-put nodes for stable learning with the adaptive learning rate. Through simulations of handwritten digit recognition and isolated-word recognition tasks, it was verified that the proposed method successfully reduced the performance dependency of MLPs on the nCE order while maintaining advantages of the nCE function.

  • PDF