• Title/Summary/Keyword: gradient descent optimization

Search Result 82, Processing Time 0.025 seconds

Parameter Learning of Dynamic Bayesian Networks using Constrained Least Square Estimation and Steepest Descent Algorithm (제약조건을 갖는 최소자승 추정기법과 최급강하 알고리즘을 이용한 동적 베이시안 네트워크의 파라미터 학습기법)

  • Cho, Hyun-Cheol;Lee, Kwon-Soon;Koo, Kyung-Wan
    • The Transactions of the Korean Institute of Electrical Engineers P
    • /
    • v.58 no.2
    • /
    • pp.164-171
    • /
    • 2009
  • This paper presents new learning algorithm of dynamic Bayesian networks (DBN) by means of constrained least square (LS) estimation algorithm and gradient descent method. First, we propose constrained LS based parameter estimation for a Markov chain (MC) model given observation data sets. Next, a gradient descent optimization is utilized for online estimation of a hidden Markov model (HMM), which is bi-linearly constructed by adding an observation variable to a MC model. We achieve numerical simulations to prove its reliability and superiority in which a series of non stationary random signal is applied for the DBN models respectively.

Improving the Training Performance of Neural Networks by using Hybrid Algorithm (하이브리드 알고리즘을 이용한 신경망의 학습성능 개선)

  • Kim, Weon-Ook;Cho, Yong-Hyun;Kim, Young-Il;Kang, In-Ku
    • The Transactions of the Korea Information Processing Society
    • /
    • v.4 no.11
    • /
    • pp.2769-2779
    • /
    • 1997
  • This Paper Proposes an efficient method for improving the training performance of the neural networks using a hybrid of conjugate gradient backpropagation algorithm and dynamic tunneling backpropagation algorithm The conjugate gradient backpropagation algorithm, which is the fast gradient algorithm, is applied for high speed optimization. The dynamic tunneling backpropagation algorithm, which is the deterministic method with tunneling phenomenon, is applied for global optimization. Conversing to the local minima by using the conjugate gradient backpropagation algorithm, the new initial point for escaping the local minima is estimated by dynamic tunneling backpropagation algorithm. The proposed method has been applied to the parity check and the pattern classification. The simulation results show that the performance of proposed method is superior to those of gradient descent backpropagtion algorithm and a hybrid of gradient descent and dynamic tunneling backpropagation algorithm, and the new algorithm converges more often to the global minima than gradient descent backpropagation algorithm.

  • PDF

An Efficient Fault-diagnosis of Digital Circuits Using Multilayer Neural Networks (다층신경망을 이용한 디지털회로의 효율적인 결함진단)

  • 조용현;박용수
    • Proceedings of the IEEK Conference
    • /
    • 1999.06a
    • /
    • pp.1033-1036
    • /
    • 1999
  • This paper proposes an efficient fault diagnosis for digital circuits using multilayer neural networks. The efficient learning algorithm is also proposed for the multilayer neural network, which is combined the steepest descent for high-speed optimization and the dynamic tunneling for global optimization. The fault-diagnosis system using the multilayer neural network of the proposed algorithm has been applied to the parity generator circuit. The simulation results shows that the proposed system is higher convergence speed and rate, in comparision with system using the backpropagation algorithm based on the gradient descent.

  • PDF

A Study on Numerical Optimization Method for Aerodynamic Design (공력설계를 위한 수치최적설계기법의 연구)

  • Jin, Xue-Song;Choi, Jae-Ho;Kim, Kwang-Yong
    • The KSFM Journal of Fluid Machinery
    • /
    • v.2 no.1 s.2
    • /
    • pp.29-34
    • /
    • 1999
  • To develop the efficient numerical optimization method for the design of an airfoil, an evaluation of various methods coupled with two-dimensional Naviev-Stokes analysis is presented. Simplex method and Hook-Jeeves method we used as direct search methods, and steepest descent method, conjugate gradient method and DFP method are used as indirect search methods and are tested to determine the search direction. To determine the moving distance, the golden section method and cubic interpolation method are tested. The finite volume method is used to discretize two-dimensional Navier-Stokes equations, and SIMPLEC algorithm is used for a velocity-pressure correction method. For the optimal design of two-dimensional airfoil, maximum thickness, maximum ordinate of camber line and chordwise position of maximum ordinate are chosen as design variables, and the ratio of drag coefficient to lift coefficient is selected as an objective function. From the results, it is found that conjugate gradient method and cubic interpolation method are the most efficient for the determination of search direction and the moving distance, respectively.

  • PDF

Self-Organizing Fuzzy Modeling Based on Hyperplane-Shaped Clusters (다차원 평면 클러스터를 이용한 자기 구성 퍼지 모델링)

  • Koh, Taek-Beom
    • Journal of Institute of Control, Robotics and Systems
    • /
    • v.7 no.12
    • /
    • pp.985-992
    • /
    • 2001
  • This paper proposes a self-organizing fuzzy modeling(SOFUM)which an create a new hyperplane shaped cluster and adjust parameters of the fuzzy model in repetition. The suggested algorithm SOFUM is composed of four steps: coarse tuning. fine tuning cluster creation and optimization of learning rates. In the coarse tuning fuzzy C-regression model(FCRM) clustering and weighted recursive least squared (WRLS) algorithm are used and in the fine tuning gradient descent algorithm is used to adjust parameters of the fuzzy model precisely. In the cluster creation, a new hyperplane shaped cluster is created by applying multiple regression to input/output data with relatively large fuzzy entropy based on parameter tunings of fuzzy model. And learning rates are optimized by utilizing meiosis-genetic algorithm in the optimization of learning rates To check the effectiveness of the suggested algorithm two examples are examined and the performance of the identified fuzzy model is demonstrated via computer simulation.

  • PDF

Learning an Artificial Neural Network Using Dynamic Particle Swarm Optimization-Backpropagation: Empirical Evaluation and Comparison

  • Devi, Swagatika;Jagadev, Alok Kumar;Patnaik, Srikanta
    • Journal of information and communication convergence engineering
    • /
    • v.13 no.2
    • /
    • pp.123-131
    • /
    • 2015
  • Training neural networks is a complex task with great importance in the field of supervised learning. In the training process, a set of input-output patterns is repeated to an artificial neural network (ANN). From those patterns weights of all the interconnections between neurons are adjusted until the specified input yields the desired output. In this paper, a new hybrid algorithm is proposed for global optimization of connection weights in an ANN. Dynamic swarms are shown to converge rapidly during the initial stages of a global search, but around the global optimum, the search process becomes very slow. In contrast, the gradient descent method can achieve faster convergence speed around the global optimum, and at the same time, the convergence accuracy can be relatively high. Therefore, the proposed hybrid algorithm combines the dynamic particle swarm optimization (DPSO) algorithm with the backpropagation (BP) algorithm, also referred to as the DPSO-BP algorithm, to train the weights of an ANN. In this paper, we intend to show the superiority (time performance and quality of solution) of the proposed hybrid algorithm (DPSO-BP) over other more standard algorithms in neural network training. The algorithms are compared using two different datasets, and the results are simulated.

Evaluation of Regression Models with various Criteria and Optimization Methods for Pollutant Load Estimations (다양한 평가 지표와 최적화 기법을 통한 오염부하 산정 회귀 모형 평가)

  • Kim, Jonggun;Lim, Kyoung Jae;Park, Youn Shik
    • Proceedings of the Korea Water Resources Association Conference
    • /
    • 2018.05a
    • /
    • pp.448-448
    • /
    • 2018
  • In this study, the regression models (Load ESTimator and eight-parameter model) were evaluated to estimate instantaneous pollutant loads under various criteria and optimization methods. As shown in the results, LOADEST commonly used in interpolating pollutant loads could not necessarily provide the best results with the automatic selected regression model. It is inferred that the various regression models in LOADEST need to be considered to find the best solution based on the characteristics of watersheds applied. The recently developed eight-parameter model integrated with Genetic Algorithm (GA) and Gradient Descent Method (GDM) were also compared with LOADEST indicating that the eight-parameter model performed better than LOADEST, but it showed different behaviors in calibration and validation. The eight-parameter model with GDM could reproduce the nitrogen loads properly outside of calibration period (validation). Furthermore, the accuracy and precision of model estimations were evaluated using various criteria (e.g., $R^2$ and gradient and constant of linear regression line). The results showed higher precisions with the $R^2$ values closed to 1.0 in LOADEST and better accuracy with the constants (in linear regression line) closed to 0.0 in the eight-parameter model with GDM. In hence, based on these finding we recommend that users need to evaluate the regression models under various criteria and calibration methods to provide the more accurate and precise results for pollutant load estimations.

  • PDF

A NONLINEAR CONJUGATE GRADIENT METHOD AND ITS GLOBAL CONVERGENCE ANALYSIS

  • CHU, AJIE;SU, YIXIAO;DU, SHOUQIANG
    • Journal of applied mathematics & informatics
    • /
    • v.34 no.1_2
    • /
    • pp.157-165
    • /
    • 2016
  • In this paper, we develop a new hybridization conjugate gradient method for solving the unconstrained optimization problem. Under mild assumptions, we get the sufficient descent property of the given method. The global convergence of the given method is also presented under the Wolfe-type line search and the general Wolfe line search. The numerical results show that the method is also efficient.

An Efficient Traning of Multilayer Neural Newtorks Using Stochastic Approximation and Conjugate Gradient Method (확률적 근사법과 공액기울기법을 이용한 다층신경망의 효율적인 학습)

  • 조용현
    • Journal of the Korean Institute of Intelligent Systems
    • /
    • v.8 no.5
    • /
    • pp.98-106
    • /
    • 1998
  • This paper proposes an efficient learning algorithm for improving the training performance of the neural network. The proposed method improves the training performance by applying the backpropagation algorithm of a global optimization method which is a hybrid of a stochastic approximation and a conjugate gradient method. The approximate initial point for f a ~gtl obal optimization is estimated first by applying the stochastic approximation, and then the conjugate gradient method, which is the fast gradient descent method, is applied for a high speed optimization. The proposed method has been applied to the parity checking and the pattern classification, and the simulation results show that the performance of the proposed method is superior to those of the conventional backpropagation and the backpropagation algorithm which is a hyhrid of the stochastic approximation and steepest descent method.

  • PDF

Cluster Analysis Algorithms Based on the Gradient Descent Procedure of a Fuzzy Objective Function

  • Rhee, Hyun-Sook;Oh, Kyung-Whan
    • Journal of Electrical Engineering and information Science
    • /
    • v.2 no.6
    • /
    • pp.191-196
    • /
    • 1997
  • Fuzzy clustering has been playing an important role in solving many problems. Fuzzy c-Means(FCM) algorithm is most frequently used for fuzzy clustering. But some fixed point of FCM algorithm, know as Tucker's counter example, is not a reasonable solution. Moreover, FCM algorithm is impossible to perform the on-line learning since it is basically a batch learning scheme. This paper presents unsupervised learning networks as an attempt to improve shortcomings of the conventional clustering algorithm. This model integrates optimization function of FCM algorithm into unsupervised learning networks. The learning rule of the proposed scheme is a result of formal derivation based on the gradient descent procedure of a fuzzy objective function. Using the result of formal derivation, two algorithms of fuzzy cluster analysis, the batch learning version and on-line learning version, are devised. They are tested on several data sets and compared with FCM. The experimental results show that the proposed algorithms find out the reasonable solution on Tucker's counter example.

  • PDF