• Title/Summary/Keyword: gradient descent optimization

Search Result 82, Processing Time 0.021 seconds

Intelligent IIR Filter based Multiple-Channel ANC Systems (지능형 IIR 필터 기반 다중 채널 ANC 시스템)

  • Cho, Hyun-Cheol;Yeo, Dae-Yeon;Lee, Young-Jin;Lee, Kwon-Soon
    • Journal of Institute of Control, Robotics and Systems
    • /
    • v.16 no.12
    • /
    • pp.1220-1225
    • /
    • 2010
  • This paper proposes a novel active noise control (ANC) approach that uses an IIR filter and neural network techniques to effectively reduce interior noise. We construct a multiple-channel IIR filter module which is a linearly augmented framework with a generic IIR model to generate a primary control signal. A three-layer perceptron neural network is employed for establishing a secondary-path model to represent air channels among noise fields. Since the IIR module and neural network are connected in series, the output of an IIR filter is transferred forward to the neural model to generate a final ANC signal. A gradient descent optimization based learning algorithm is analytically derived for the optimal selection of the ANC parameter vectors. Moreover, re-estimation of partial parameter vectors in the ANC system is proposed for online learning. Lastly, we present the results of a numerical study to test our ANC methodology with realistic interior noise measurement obtained from Korean railway trains.

Development of the Adaptive PPF Controller for the Vibration Syppression of Smart Structures (지능구조물 제어를 위한 적응형 PPF 제어기의 개발)

  • Lee, Seung-Bum;Heo, Seok;Kwak, Moom Ku
    • Proceedings of the Korean Society for Noise and Vibration Engineering Conference
    • /
    • 2001.05a
    • /
    • pp.302-307
    • /
    • 2001
  • This research is concerned with the development of a real-time adaptive PPF controller for the active vibration suppression of smart structure. In general, the tuning of the PPF controller is carried out off-line. In this research, the real-time learning algorithm is developed to find the optimal filter frequency of the PPF controller in real time and the efficacy of the algorithm is proved by implementing it in real time. To this end, the adaptive algorithm is developed by applying the gradient descent method to the predefined performance index, which is similar to the method used popularly in the optimization and neural network controller design. The experiment was carried out to verify the validity of the adaptive PPF controller developed in this research. The experimental results showed that adaptive PPF controller is effective for active vibration control of the structure which is excited by either impact or harmonic disturbance. The filter frequency of the PPF controller can be tuned in a very short period of time thus proving the efficiency of the adaptive PPF controller.

  • PDF

Hybrid Adaptive Feedforward Control System Against State and Input Disturbances (시스템 상태 및 입력 외란을 고려한 하이브리드 방식의 적응형 피드포워드 제어시스템)

  • Kim, Jun-Su;Cho, Hyun-Cheol;Kim, Gwan-Hyung;Ha, Hong-Gon;Lee, Hyung-Ki
    • Journal of Institute of Control, Robotics and Systems
    • /
    • v.18 no.3
    • /
    • pp.237-242
    • /
    • 2012
  • AFC (Adaptive Feedforward Control) is significantly employed for improving control performance of dynamic systems particularly involving periodic disturbance signals in engineering fields. This paper presents a novel hybrid AFC approach for discrete-time systems with multiple disturbances in terms of control input and state variables. The proposed AFC mechanism is hierarchically composed of a conventional feedforward control framework and PID auxiliary control configuration in parallel. The former is generic to decrease periodic disturbance excited to control actuators and the latter is additionally constructed to overcome control deterioration due to time-varying uncertainty under given systems. We carry out numerical simulation to test reliability of our proposed hybrid AFC system and compare its control performance to a well-known conventional AFC method with respect to time and frequency domains for proving of its superiority.

Searching a global optimum by stochastic perturbation in error back-propagation algorithm (오류 역전파 학습에서 확률적 가중치 교란에 의한 전역적 최적해의 탐색)

  • 김삼근;민창우;김명원
    • Journal of the Korean Institute of Telematics and Electronics C
    • /
    • v.35C no.3
    • /
    • pp.79-89
    • /
    • 1998
  • The Error Back-Propagation(EBP) algorithm is widely applied to train a multi-layer perceptron, which is a neural network model frequently used to solve complex problems such as pattern recognition, adaptive control, and global optimization. However, the EBP is basically a gradient descent method, which may get stuck in a local minimum, leading to failure in finding the globally optimal solution. Moreover, a multi-layer perceptron suffers from locking a systematic determination of the network structure appropriate for a given problem. It is usually the case to determine the number of hidden nodes by trial and error. In this paper, we propose a new algorithm to efficiently train a multi-layer perceptron. OUr algorithm uses stochastic perturbation in the weight space to effectively escape from local minima in multi-layer perceptron learning. Stochastic perturbation probabilistically re-initializes weights associated with hidden nodes to escape a local minimum if the probabilistically re-initializes weights associated with hidden nodes to escape a local minimum if the EGP learning gets stuck to it. Addition of new hidden nodes also can be viewed asa special case of stochastic perturbation. Using stochastic perturbation we can solve the local minima problem and the network structure design in a unified way. The results of our experiments with several benchmark test problems including theparity problem, the two-spirals problem, andthe credit-screening data show that our algorithm is very efficient.

  • PDF

Pragmatic Assessment of Optimizers in Deep Learning

  • Ajeet K. Jain;PVRD Prasad Rao ;K. Venkatesh Sharma
    • International Journal of Computer Science & Network Security
    • /
    • v.23 no.10
    • /
    • pp.115-128
    • /
    • 2023
  • Deep learning has been incorporating various optimization techniques motivated by new pragmatic optimizing algorithm advancements and their usage has a central role in Machine learning. In recent past, new avatars of various optimizers are being put into practice and their suitability and applicability has been reported on various domains. The resurgence of novelty starts from Stochastic Gradient Descent to convex and non-convex and derivative-free approaches. In the contemporary of these horizons of optimizers, choosing a best-fit or appropriate optimizer is an important consideration in deep learning theme as these working-horse engines determines the final performance predicted by the model. Moreover with increasing number of deep layers tantamount higher complexity with hyper-parameter tuning and consequently need to delve for a befitting optimizer. We empirically examine most popular and widely used optimizers on various data sets and networks-like MNIST and GAN plus others. The pragmatic comparison focuses on their similarities, differences and possibilities of their suitability for a given application. Additionally, the recent optimizer variants are highlighted with their subtlety. The article emphasizes on their critical role and pinpoints buttress options while choosing among them.

Anisotropic Total Variation Denoising Technique for Low-Dose Cone-Beam Computed Tomography Imaging

  • Lee, Ho;Yoon, Jeongmin;Lee, Eungman
    • Progress in Medical Physics
    • /
    • v.29 no.4
    • /
    • pp.150-156
    • /
    • 2018
  • This study aims to develop an improved Feldkamp-Davis-Kress (FDK) reconstruction algorithm using anisotropic total variation (ATV) minimization to enhance the image quality of low-dose cone-beam computed tomography (CBCT). The algorithm first applies a filter that integrates the Shepp-Logan filter into a cosine window function on all projections for impulse noise removal. A total variation objective function with anisotropic penalty is then minimized to enhance the difference between the real structure and noise using the steepest gradient descent optimization with adaptive step sizes. The preserving parameter to adjust the separation between the noise-free and noisy areas is determined by calculating the cumulative distribution function of the gradient magnitude of the filtered image obtained by the application of the filtering operation on each projection. With these minimized ATV projections, voxel-driven backprojection is finally performed to generate the reconstructed images. The performance of the proposed algorithm was evaluated with the catphan503 phantom dataset acquired with the use of a low-dose protocol. Qualitative and quantitative analyses showed that the proposed ATV minimization provides enhanced CBCT reconstruction images compared with those generated by the conventional FDK algorithm, with a higher contrast-to-noise ratio (CNR), lower root-mean-square-error, and higher correlation. The proposed algorithm not only leads to a potential imaging dose reduction in repeated CBCT scans via lower mA levels, but also elicits high CNR values by removing noisy corrupted areas and by avoiding the heavy penalization of striking features.

The Optimization of Ensembles for Bankruptcy Prediction (기업부도 예측 앙상블 모형의 최적화)

  • Myoung Jong Kim;Woo Seob Yun
    • Information Systems Review
    • /
    • v.24 no.1
    • /
    • pp.39-57
    • /
    • 2022
  • This paper proposes the GMOPTBoost algorithm to improve the performance of the AdaBoost algorithm for bankruptcy prediction in which class imbalance problem is inherent. AdaBoost algorithm has the advantage of providing a robust learning opportunity for misclassified samples. However, there is a limitation in addressing class imbalance problem because the concept of arithmetic mean accuracy is embedded in AdaBoost algorithm. GMOPTBoost can optimize the geometric mean accuracy and effectively solve the category imbalance problem by applying Gaussian gradient descent. The samples are constructed according to the following two phases. First, five class imbalance datasets are constructed to verify the effect of the class imbalance problem on the performance of the prediction model and the performance improvement effect of GMOPTBoost. Second, class balanced data are constituted through data sampling techniques to verify the performance improvement effect of GMOPTBoost. The main results of 30 times of cross-validation analyzes are as follows. First, the class imbalance problem degrades the performance of ensembles. Second, GMOPTBoost contributes to performance improvements of AdaBoost ensembles trained on imbalanced datasets. Third, Data sampling techniques have a positive impact on performance improvement. Finally, GMOPTBoost contributes to significant performance improvement of AdaBoost ensembles trained on balanced datasets.

Deep Learning Architectures and Applications (딥러닝의 모형과 응용사례)

  • Ahn, SungMahn
    • Journal of Intelligence and Information Systems
    • /
    • v.22 no.2
    • /
    • pp.127-142
    • /
    • 2016
  • Deep learning model is a kind of neural networks that allows multiple hidden layers. There are various deep learning architectures such as convolutional neural networks, deep belief networks and recurrent neural networks. Those have been applied to fields like computer vision, automatic speech recognition, natural language processing, audio recognition and bioinformatics where they have been shown to produce state-of-the-art results on various tasks. Among those architectures, convolutional neural networks and recurrent neural networks are classified as the supervised learning model. And in recent years, those supervised learning models have gained more popularity than unsupervised learning models such as deep belief networks, because supervised learning models have shown fashionable applications in such fields mentioned above. Deep learning models can be trained with backpropagation algorithm. Backpropagation is an abbreviation for "backward propagation of errors" and a common method of training artificial neural networks used in conjunction with an optimization method such as gradient descent. The method calculates the gradient of an error function with respect to all the weights in the network. The gradient is fed to the optimization method which in turn uses it to update the weights, in an attempt to minimize the error function. Convolutional neural networks use a special architecture which is particularly well-adapted to classify images. Using this architecture makes convolutional networks fast to train. This, in turn, helps us train deep, muti-layer networks, which are very good at classifying images. These days, deep convolutional networks are used in most neural networks for image recognition. Convolutional neural networks use three basic ideas: local receptive fields, shared weights, and pooling. By local receptive fields, we mean that each neuron in the first(or any) hidden layer will be connected to a small region of the input(or previous layer's) neurons. Shared weights mean that we're going to use the same weights and bias for each of the local receptive field. This means that all the neurons in the hidden layer detect exactly the same feature, just at different locations in the input image. In addition to the convolutional layers just described, convolutional neural networks also contain pooling layers. Pooling layers are usually used immediately after convolutional layers. What the pooling layers do is to simplify the information in the output from the convolutional layer. Recent convolutional network architectures have 10 to 20 hidden layers and billions of connections between units. Training deep learning networks has taken weeks several years ago, but thanks to progress in GPU and algorithm enhancement, training time has reduced to several hours. Neural networks with time-varying behavior are known as recurrent neural networks or RNNs. A recurrent neural network is a class of artificial neural network where connections between units form a directed cycle. This creates an internal state of the network which allows it to exhibit dynamic temporal behavior. Unlike feedforward neural networks, RNNs can use their internal memory to process arbitrary sequences of inputs. Early RNN models turned out to be very difficult to train, harder even than deep feedforward networks. The reason is the unstable gradient problem such as vanishing gradient and exploding gradient. The gradient can get smaller and smaller as it is propagated back through layers. This makes learning in early layers extremely slow. The problem actually gets worse in RNNs, since gradients aren't just propagated backward through layers, they're propagated backward through time. If the network runs for a long time, that can make the gradient extremely unstable and hard to learn from. It has been possible to incorporate an idea known as long short-term memory units (LSTMs) into RNNs. LSTMs make it much easier to get good results when training RNNs, and many recent papers make use of LSTMs or related ideas.

Hierarchical Particle Swarm Optimization for Multi UAV Waypoints Planning Under Various Threats (다양한 위협 하에서 복수 무인기의 경로점 계획을 위한 계층적 입자 군집 최적화)

  • Chung, Wonmo;Kim, Myunggun;Lee, Sanha;Lee, Sang-Pill;Park, Chun-Shin;Son, Hungsun
    • Journal of the Korean Society for Aeronautical & Space Sciences
    • /
    • v.50 no.6
    • /
    • pp.385-391
    • /
    • 2022
  • This paper presents to develop a path planning algorithm combining gradient descent-based path planning (GBPP) and particle swarm optimization (PSO) for considering prohibited flight areas, terrain information, and characteristics of fixed-wing unmmaned aerial vehicle (UAV) in 3D space. Path can be generated fast using GBPP, but it is often happened that an unsafe path can be generated by converging to a local minimum depending on the initial path. Bio-inspired swarm intelligence algorithms, such as Genetic algorithm (GA) and PSO, can avoid the local minima problem by sampling several paths. However, if the number of optimal variable increases due to an increase in the number of UAVs and waypoints, it requires heavy computation time and efforts due to increasing the number of particles accordingly. To solve the disadvantages of the two algorithms, hierarchical path planning algorithm associated with hierarchical particle swarm optimization (HPSO) is developed by defining the initial path, which is the input of GBPP, as two variables including particles variables. Feasibility of the proposed algorithm is verified by software-in-the-loop simulation (SILS) of flight control computer (FCC) for UAVs.

Analysis of Microwave Inverse Scattering Using the Broadband Electromagnetic Waves (광대역 전자파를 이용한 역산란 해석 연구)

  • Lee Jung-Hoon;Chung Young-Seek;So Joon-Ho;Kim Junyeon;Jang Won
    • The Journal of Korean Institute of Electromagnetic Engineering and Science
    • /
    • v.17 no.2 s.105
    • /
    • pp.158-164
    • /
    • 2006
  • In this paper, we proposed a new algorithm of the inverse scattering for the reconstruction of unknown dielectric scatterers using the finite-difference time-domain method and the design sensitivity analysis. We introduced the design sensitivity analysis based on the gradient information for the fast convergence of the reconstruction. By introducing the adjoint variable method for the efficient calculation, we derived the adjoint variable equation. As an optimal algorithm, we used the steepest descent method and reconstructed the dielectric targets using the iterative estimation. To verify our algorithm, we will show the numerical examples for the two-dimensional $TM^2$ cases.