• Title/Summary/Keyword: Gradient descent

Search Result 339, Processing Time 0.022 seconds

Learning algorithms for big data logistic regression on RHIPE platform (RHIPE 플랫폼에서 빅데이터 로지스틱 회귀를 위한 학습 알고리즘)

  • Jung, Byung Ho;Lim, Dong Hoon
    • Journal of the Korean Data and Information Science Society
    • /
    • v.27 no.4
    • /
    • pp.911-923
    • /
    • 2016
  • Machine learning becomes increasingly important in the big data era. Logistic regression is a type of classification in machine leaning, and has been widely used in various fields, including medicine, economics, marketing, and social sciences. Rhipe that integrates R and Hadoop environment, has not been discussed by many researchers owing to the difficulty of its installation and MapReduce implementation. In this paper, we present the MapReduce implementation of Gradient Descent algorithm and Newton-Raphson algorithm for logistic regression using Rhipe. The Newton-Raphson algorithm does not require a learning rate, while Gradient Descent algorithm needs to manually pick a learning rate. We choose the learning rate by performing the mixed procedure of grid search and binary search for processing big data efficiently. In the performance study, our Newton-Raphson algorithm outpeforms Gradient Descent algorithm in all the tested data.

Direct Gradient Descent Control and Sontag's Formula on Asymptotic Stability of General Nonlinear Control System

  • Naiborhu J.;Nababan S. M.;Saragih R.;Pranoto I.
    • International Journal of Control, Automation, and Systems
    • /
    • v.3 no.2
    • /
    • pp.244-251
    • /
    • 2005
  • In this paper, we study the problem of stabilizing a general nonlinear control system by means of gradient descent control method which is a dynamic feedback control law. In this method, the general nonlinear control system can be considered as an affine nonlinear control systems. Then by using Sontag's formula we investigate the stability (asymptotic) of the general nonlinear control system.

Optimal Learning Rates in Gradient Descent Training of Multilayer Perceptrons (다층퍼셉트론의 강하 학습을 위한 최적 학습률)

  • 오상훈
    • The Journal of the Korea Contents Association
    • /
    • v.4 no.3
    • /
    • pp.99-105
    • /
    • 2004
  • This paper proposes optimal learning rates in the gradient descent training of multilayer perceptrons, which are a separate learning rate for weights associated with each neuron and a separate one for assigning virtual hidden targets associated with each training pattern Effectiveness of the proposed error function was demonstrated for a handwritten digit recognition and an isolated-word recognition tasks and very fast learning convergence was obtained.

  • PDF

A new optimization method for improving the performance of neural networks for optimization (최적화용 신경망의 성능개선을 위한 새로운 최적화 기법)

  • 조영현
    • Journal of the Korean Institute of Telematics and Electronics C
    • /
    • v.34C no.12
    • /
    • pp.61-69
    • /
    • 1997
  • This paper proposes a new method for improving the performances of the neural network for optimization using a hyubrid of gradient descent method and dynamic tunneling system. The update rule of gradient descent method, which has the fast convergence characteristic, is applied for high-speed optimization. The update rule of dynamic tunneling system, which is the deterministic method with a tunneling phenomenon, is applied for global optimization. Having converged to the for escaping the local minima by applying the dynamic tunneling system. The proposed method has been applied to the travelling salesman problems and the optimal task partition problems to evaluate to that of hopfield model using the update rule of gradient descent method.

  • PDF

Parameter Learning of Dynamic Bayesian Networks using Constrained Least Square Estimation and Steepest Descent Algorithm (제약조건을 갖는 최소자승 추정기법과 최급강하 알고리즘을 이용한 동적 베이시안 네트워크의 파라미터 학습기법)

  • Cho, Hyun-Cheol;Lee, Kwon-Soon;Koo, Kyung-Wan
    • The Transactions of the Korean Institute of Electrical Engineers P
    • /
    • v.58 no.2
    • /
    • pp.164-171
    • /
    • 2009
  • This paper presents new learning algorithm of dynamic Bayesian networks (DBN) by means of constrained least square (LS) estimation algorithm and gradient descent method. First, we propose constrained LS based parameter estimation for a Markov chain (MC) model given observation data sets. Next, a gradient descent optimization is utilized for online estimation of a hidden Markov model (HMM), which is bi-linearly constructed by adding an observation variable to a MC model. We achieve numerical simulations to prove its reliability and superiority in which a series of non stationary random signal is applied for the DBN models respectively.

Performance Comparison of the Optimizers in a Faster R-CNN Model for Object Detection of Metaphase Chromosomes (중기 염색체 객체 검출을 위한 Faster R-CNN 모델의 최적화기 성능 비교)

  • Jung, Wonseok;Lee, Byeong-Soo;Seo, Jeongwook
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.23 no.11
    • /
    • pp.1357-1363
    • /
    • 2019
  • In this paper, we compares the performance of the gredient descent optimizers of the Faster Region-based Convolutional Neural Network (R-CNN) model for the chromosome object detection in digital images composed of human metaphase chromosomes. In faster R-CNN, the gradient descent optimizer is used to minimize the objective function of the region proposal network (RPN) module and the classification score and bounding box regression blocks. The gradient descent optimizer. Through performance comparisons among these four gradient descent optimizers in our experiments, we found that the Adamax optimizer could achieve the mean average precision (mAP) of about 52% when considering faster R-CNN with a base network, VGG16. In case of faster R-CNN with a base network, ResNet50, the Adadelta optimizer could achieve the mAP of about 58%.

Fuzzy Modeling based on FCM Clustering Algorithm (FCM 클러스터링 알고리즘에 기초한 퍼지 모델링)

  • 윤기찬;오성권
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 2000.10a
    • /
    • pp.373-373
    • /
    • 2000
  • In this paper, we propose a fuzzy modeling algorithm which divides the input space more efficiently than convention methods by taking into consideration correlations between components of sample data. The proposed fuzzy modeling algorithm consists of two steps: coarse tuning, which determines consequent parameters approximately using FCRM clustering method, and fine tuning, which adjusts the premise and consequent parameters more precisely by gradient descent algorithm. To evaluate the performance of the proposed fuzzy mode, we use the numerical data of nonlinear function.

  • PDF

Intracranial Hemorrhagic Lesion Feature Extraction System Of Using Wavelet Transform and LMBP (웨이블렛 변환과 LMBP를 이용한 대뇌출혈성 병변 인식 시스템)

  • 정유정;정채영
    • Proceedings of the Korean Information Science Society Conference
    • /
    • 2002.10d
    • /
    • pp.625-627
    • /
    • 2002
  • 본 논문에서는 의료영상 인식 기술 중 인식률이 뛰어나고 신뢰성 있는 대뇌출혈성 병변인식 시스템을 구현하기 위하여 신호처리 분야에서 주로 사용되는 Wavelet 변환과 신경망 기법을 이용하고자 한다. 그러나 신경망 기법에서 주로 사용되는 비선형 최적화기법인 Gradient descent BP는 최적화 문제점을 해결하기에는 수렴속도가 느리기 때문에 적합하지 않다. 따라서 본 논문에서는 기존 Gradient descent BP를 보완한 Levenberg-Marquardt Back-Propagation을 대뇌출혈성 병변인식에 적용하여 구현함으로써 총 50개의 패턴 중 45개의 영상이 인식에 성공하였고 전체 평균 인식률은 각각 90%와 87%의 인식률을 보였다.

  • PDF

Improving the Training Performance of Neural Networks by using Hybrid Algorithm (하이브리드 알고리즘을 이용한 신경망의 학습성능 개선)

  • Kim, Weon-Ook;Cho, Yong-Hyun;Kim, Young-Il;Kang, In-Ku
    • The Transactions of the Korea Information Processing Society
    • /
    • v.4 no.11
    • /
    • pp.2769-2779
    • /
    • 1997
  • This Paper Proposes an efficient method for improving the training performance of the neural networks using a hybrid of conjugate gradient backpropagation algorithm and dynamic tunneling backpropagation algorithm The conjugate gradient backpropagation algorithm, which is the fast gradient algorithm, is applied for high speed optimization. The dynamic tunneling backpropagation algorithm, which is the deterministic method with tunneling phenomenon, is applied for global optimization. Conversing to the local minima by using the conjugate gradient backpropagation algorithm, the new initial point for escaping the local minima is estimated by dynamic tunneling backpropagation algorithm. The proposed method has been applied to the parity check and the pattern classification. The simulation results show that the performance of proposed method is superior to those of gradient descent backpropagtion algorithm and a hybrid of gradient descent and dynamic tunneling backpropagation algorithm, and the new algorithm converges more often to the global minima than gradient descent backpropagation algorithm.

  • PDF

Hybrid Fuzzy Adaptive Control of LEGO Robots

  • Vaseak, Jan;Miklos, Marian
    • International Journal of Fuzzy Logic and Intelligent Systems
    • /
    • v.2 no.1
    • /
    • pp.65-69
    • /
    • 2002
  • The main drawback of “classical”fuzzy systems is the inability to design and maintain their database. To overcome this disadvantage many types of extensions adding the adaptivity property to those systems were designed. This paper deals with one of them a new hybrid adaptation structure, called gradient-incremental adaptive fuzzy controller connecting gradient-descent methods with the so-called self-organizing fuzzy logic controller designed by Procyk and Mamdani. The aim is to incorporate the advantages of both Principles. This controller was implemented and tested on the system of LEGO robots. The results and comparison to a ‘classical’(non-adaptive) fuzzy controller designed by a human operator are also shown here.