• Title/Summary/Keyword: Gradient Descent Learning

Search Result 152, Processing Time 0.029 seconds

Multi-layer Neural Network with Hybrid Learning Rules for Improved Robust Capability (Robustness를 형성시키기 위한 Hybrid 학습법칙을 갖는 다층구조 신경회로망)

  • 정동규;이수영
    • Journal of the Korean Institute of Telematics and Electronics B
    • /
    • v.31B no.8
    • /
    • pp.211-218
    • /
    • 1994
  • In this paper we develope a hybrid learning rule to improve the robustness of multi-layer Perceptions. In most neural networks the activation of a neuron is deternined by a nonlinear transformation of the weighted sum of inputs to the neurons. Investigating the behaviour of activations of hidden layer neurons a new learning algorithm is developed for improved robustness for multi-layer Perceptrons. Unlike other methods which reduce the network complexity by putting restrictions on synaptic weights our method based on error-backpropagation increases the complexity of the underlying proplem by imposing it saturation requirement on hidden layer neurons. We also found that the additional gradient-descent term for the requirement corresponds to the Hebbian rule and our algorithm incorporates the Hebbian learning rule into the error back-propagation rule. Computer simulation demonstrates fast learning convergence as well as improved robustness for classification and hetero-association of patterns.

  • PDF

Stable Tracking Control to a Non-linear Process Via Neural Network Model

  • Zhai, Yujia
    • Journal of the Korea Convergence Society
    • /
    • v.5 no.4
    • /
    • pp.163-169
    • /
    • 2014
  • A stable neural network control scheme for unknown non-linear systems is developed in this paper. While the control variable is optimised to minimize the performance index, convergence of the index is guaranteed asymptotically stable by a Lyapnov control law. The optimization is achieved using a gradient descent searching algorithm and is consequently slow. A fast convergence algorithm using an adaptive learning rate is employed to speed up the convergence. Application of the stable control to a single input single output (SISO) non-linear system is simulated. The satisfactory control performance is obtained.

Optimization of Block-based Evolvable Neural Network using the Genetic Algorithm (유전자 알고리즘을 이용한 블록 기반 진화신경망의 최적화)

  • 문상우;공성곤
    • Proceedings of the IEEK Conference
    • /
    • 1999.06a
    • /
    • pp.460-463
    • /
    • 1999
  • In this paper, we proposed an block-based evolvable neural network(BENN). The BENN can optimize it's structure and weights simultaneously. It can be easily implemented by FPGA whose connection and internal functionality can be reconfigured. To solve the local minima problem that is caused gradient descent learning algorithm, genetic algorithms are applied for optimizing the proposed evolvable neural network model.

  • PDF

An Efficient Traning of Multilayer Neural Newtorks Using Stochastic Approximation and Conjugate Gradient Method (확률적 근사법과 공액기울기법을 이용한 다층신경망의 효율적인 학습)

  • 조용현
    • Journal of the Korean Institute of Intelligent Systems
    • /
    • v.8 no.5
    • /
    • pp.98-106
    • /
    • 1998
  • This paper proposes an efficient learning algorithm for improving the training performance of the neural network. The proposed method improves the training performance by applying the backpropagation algorithm of a global optimization method which is a hybrid of a stochastic approximation and a conjugate gradient method. The approximate initial point for f a ~gtl obal optimization is estimated first by applying the stochastic approximation, and then the conjugate gradient method, which is the fast gradient descent method, is applied for a high speed optimization. The proposed method has been applied to the parity checking and the pattern classification, and the simulation results show that the performance of the proposed method is superior to those of the conventional backpropagation and the backpropagation algorithm which is a hyhrid of the stochastic approximation and steepest descent method.

  • PDF

On Learning of HMM-Net Classifiers Using Hybrid Methods (하이브리드법에 의한 HMM-Net 분류기의 학습)

  • 김상운;신성효
    • Proceedings of the IEEK Conference
    • /
    • 1998.10a
    • /
    • pp.1273-1276
    • /
    • 1998
  • The HMM-Net is an architecture for a neural network that implements a hidden Markov model (HMM). The architecture is developed for the purpose of combining the discriminant power of neural networks with the time-domain modeling capability of HMMs. Criteria used for learning HMM-Net classifiers are maximum likelihood (ML), maximum mutual information (MMI), and minimization of mean squared error(MMSE). In this paper we propose an efficient learning method of HMM-Net classifiers using hybrid criteria, ML/MMSE and MMI/MMSE, and report the results of an experimental study comparing the performance of HMM-Net classifiers trained by the gradient descent algorithm with the above criteria. Experimental results for the isolated numeric digits from /0/ to /9/ show that the performance of the proposed method is better than the others in the respects of learning and recognition rates.

  • PDF

Fuzzy logic control of a planar parallel manipulator using multi learning algorithm (다중 학습 알고리듬을 이용한 평면형 병렬 매니퓰레이터의 Fuzzy 논리 제어)

  • Song, Nak-Yun;Cho, Whang
    • Journal of Institute of Control, Robotics and Systems
    • /
    • v.5 no.8
    • /
    • pp.914-922
    • /
    • 1999
  • A study on the improvement of tracking performance of a 3 DOF planar parallel manipulator is performed. A class of adaptive tracking control sheme is designed using self tuning adaptive fuzzy logic control theory. This control sheme is composed of three classical PD controller and a multi learning type self tuning adaptive fuzzy logic controller set. PD controller is tuned roughly by manual setting a priori and fuzzy logic controller is tuned precisely by the gradient descent method for a global solution during run-time, so the proposed control scheme is tuned more rapidly and precisely than the single learning type self tuning adaptive fuzzy logic control sheme for a local solution. The control performance of the proposed algorithm is verified through experiments.

  • PDF

An efficient learning method of HMM-Net classifiers (HMM-Net 분류기의 효율적인 학습법)

  • 김상운;김탁령
    • Proceedings of the IEEK Conference
    • /
    • 1998.06a
    • /
    • pp.933-935
    • /
    • 1998
  • The HMM-Net is an architecture for a neural network that implements a hidden markov model (HMM). The architecture is developed for the purpose of combining the discriminant power of neural networks with the time-domain modeling capability of HMMs. Criteria used for learning HMM-Net classifiers are maximum likelihood(ML) and minimization of mean squared error(MMSE). In this paper we propose an efficient learning method of HMM_Net classifiers using a ML-MMSE hybrid criterion and report the results of an experimental study comparing the performance of HMM_Net classifiers trained by the gradient descent algorithm with the above criteria. Experimental results for the isolated numeric digits from /0/ to /9/ show that the performance of the proposed method is better than the others in the repects of learning and recognition rates.

  • PDF

A study on time-varying control of learning parameters in neural networks (신경망 학습 변수의 시변 제어에 관한 연구)

  • 박종철;원상철;최한고
    • Proceedings of the Korea Institute of Convergence Signal Processing
    • /
    • 2000.12a
    • /
    • pp.201-204
    • /
    • 2000
  • This paper describes a study on the time-varying control of parameters in learning of the neural network. Elman recurrent neural network (RNN) is used to implement the control of parameters. The parameters of learning and momentum rates In the error backpropagation algorithm ate updated at every iteration using fuzzy rules based on performance index. In addition, the gain and slope of the neuron's activation function are also considered time-varying parameters. These function parameters are updated using the gradient descent algorithm. Simulation results show that the auto-tuned learning algorithm results in faster convergence and lower system error than regular backpropagation in the system identification.

  • PDF

MIMO Fuzzy Reasoning Method using Learning Ability (학습기능을 사용한 MIMO 퍼지추론 방식)

  • Park, Jin-Hyun;Lee, Tae-Hwan;Choi, Young-Kiu
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2008.10a
    • /
    • pp.175-178
    • /
    • 2008
  • Z. Cao had proposed NFRM(new fuzzy reasoning method) which infers in detail using relation matrix. In spite of the small inference rules, it shows good performance than mamdani's fuzzy inference method. But the most of fuzzy systems are difficult to make fuzzy inference rules in the case of MIMO system. The past days, We had proposed the MIMO fuzzy inference which had extended a Z. Cao's fuzzy inference to handle MIMO system. But many times and effort needed to determine the relation matrix elements of MIMO fuzzy inference by heuristic and trial and error method in order to improve inference performances. In this paper, we propose a MIMO fuzzy inference method with the learning ability witch is used a gradient descent method in order to improve the performances. Through the computer simulation studies for the inverse kinematics problem of 2-axis robot, we show that proposed inference method using a gradient descent method has good performances.

  • PDF

Face Recognition Based on Improved Fuzzy RBF Neural Network for Smar t Device

  • Lee, Eung-Joo
    • Journal of Korea Multimedia Society
    • /
    • v.16 no.11
    • /
    • pp.1338-1347
    • /
    • 2013
  • Face recognition is a science of automatically identifying individuals based their unique facial features. In order to avoid overfitting and reduce the computational reduce the computational burden, a new face recognition algorithm using PCA-fisher linear discriminant (PCA-FLD) and fuzzy radial basis function neural network (RBFNN) is proposed in this paper. First, face features are extracted by the principal component analysis (PCA) method. Then, the extracted features are further processed by the Fisher's linear discriminant technique to acquire lower-dimensional discriminant patterns, the processed features will be considered as the input of the fuzzy RBFNN. As a widely applied algorithm in fuzzy RBF neural network, BP learning algorithm has the low rate of convergence, therefore, an improved learning algorithm based on Levenberg-Marquart (L-M) for fuzzy RBF neural network is introduced in this paper, which combined the Gradient Descent algorithm with the Gauss-Newton algorithm. Experimental results on the ORL face database demonstrate that the proposed algorithm has satisfactory performance and high recognition rate.