• Title/Summary/Keyword: optimization algorithms

Search Result 1,702, Processing Time 0.03 seconds

Henry gas solubility optimization for control of a nuclear reactor: A case study

  • Mousakazemi, Seyed Mohammad Hossein
    • Nuclear Engineering and Technology
    • /
    • v.54 no.3
    • /
    • pp.940-947
    • /
    • 2022
  • Meta-heuristic algorithms have found their place in optimization problems. Henry gas solubility optimization (HGSO) is one of the newest population-based algorithms. This algorithm is inspired by Henry's law of physics. To evaluate the performance of a new algorithm, it must be used in various problems. On the other hand, the optimization of the proportional-integral-derivative (PID) gains for load-following of a nuclear power plant (NPP) is a good challenge to assess the performance of HGSO. Accordingly, the power control of a pressurized water reactor (PWR) is targeted, based on the point kinetics model with six groups of delayed-neutron precursors. In any optimization problem based on meta-heuristic algorithms, an efficient objective function is required. Therefore, the integral of the time-weighted square error (ITSE) performance index is utilized as the objective (cost) function of HGSO, which is constrained by a stability criterion in steady-state operations. A Lyapunov approach guarantees this stability. The results show that this method provides superior results compared to an empirically tuned PID controller with the least error. It also achieves good accuracy compared to an established GA-tuned PID controller.

Comparative Analysis of Optimization Algorithms and the Effects of Coupling Hedging Rules in Reservoir Operations

  • Kim, Gi Joo;Kim, Young-Oh
    • Proceedings of the Korea Water Resources Association Conference
    • /
    • 2021.06a
    • /
    • pp.206-206
    • /
    • 2021
  • The necessity for appropriate management of water resources infrastructures such as reservoirs, levees, and dikes is increasing due to unexpected hydro-climate irregularities and rising water demands. To meet this need, past studies have focused on advancing theoretical optimization algorithms such as nonlinear programming, dynamic programming (DP), and genetic programming. Yet, the optimally derived theoretical solutions are limited to be directly implemented in making release decisions in the real-world systems for a variety of reasons. This study first aims to comparatively analyze the two prominent optimization methods, DP and evolutionary multi-objective direct policy search (EMODPS), under historical inflow series using K-fold cross validation. A total of six optimization models are formed each with a specific formulation. Then, one of the optimization models was coupled with the actual zone-based hedging rule that has been adopted in practice. The proposed methodology was applied to Boryeong Dam located in South Korea with conflicting objectives between supply and demand. As a result, the EMODPS models demonstrated a better performance than the DP models in terms of proximity to the ideal. Moreover, the incorporation of the real-world policy with the optimal solutions improved in all indices in terms of the supply side, while widening the range of the trade-off between frequency and magnitude measured in the sides of demand. The results from this study once again highlight the necessity of closing the gap between the theoretical solutions with the real-world implementable policies.

  • PDF

A new hybrid meta-heuristic for structural design: ranked particles optimization

  • Kaveh, A.;Nasrollahi, A.
    • Structural Engineering and Mechanics
    • /
    • v.52 no.2
    • /
    • pp.405-426
    • /
    • 2014
  • In this paper, a new meta-heuristic algorithm named Ranked Particles Optimization (RPO), is presented. This algorithm is not inspired from natural or physical phenomena. However, it is based on numerous researches in the field of meta-heuristic optimization algorithms. In this algorithm, like other meta-heuristic algorithms, optimization process starts with by producing a population of random solutions, Particles, located in the feasible search space. In the next step, cost functions corresponding to all random particles are evaluated and some of those having minimum cost functions are stored. These particles are ranked and their weighted average is calculated and named Ranked Center. New solutions are produced by moving each particle along its previous motion, the ranked center, and the best particle found thus far. The robustness of this algorithm is verified by solving some mathematical and structural optimization problems. Simplicity of implementation and reaching to desired solution are two main characteristics of this algorithm.

Physics-based Surrogate Optimization of Francis Turbine Runner Blades, Using Mesh Adaptive Direct Search and Evolutionary Algorithms

  • Bahrami, Salman;Tribes, Christophe;von Fellenberg, Sven;Vu, Thi C.;Guibault, Francois
    • International Journal of Fluid Machinery and Systems
    • /
    • v.8 no.3
    • /
    • pp.209-219
    • /
    • 2015
  • A robust multi-fidelity optimization methodology has been developed, focusing on efficiently handling industrial runner design of hydraulic Francis turbines. The computational task is split between low- and high-fidelity phases in order to properly balance the CFD cost and required accuracy in different design stages. In the low-fidelity phase, a physics-based surrogate optimization loop manages a large number of iterative optimization evaluations. Two derivative-free optimization methods use an inviscid flow solver as a physics-based surrogate to obtain the main characteristics of a good design in a relatively fast iterative process. The case study of a runner design for a low-head Francis turbine indicates advantages of integrating two derivative-free optimization algorithms with different local- and global search capabilities.

Satellite Customer Assignment: A Comparative Study of Genetic Algorithm and Ant Colony Optimization

  • Kim, Sung-Soo;Kim, Hyoung-Joong;Mani, V.
    • Journal of Ubiquitous Convergence Technology
    • /
    • v.2 no.1
    • /
    • pp.40-50
    • /
    • 2008
  • The problem of assigning customers to satellite channels is a difficult combinatorial optimization problem and is NP-complete. For this combinatorial optimization problem, standard optimization methods take a large computation time and so genetic algorithms (GA) and ant colony optimization (ACO) can be used to obtain the best and/or optimal assignment of customers to satellite channels. In this paper, we present a comparative study of GA and ACO to this problem. Various issues related to genetic algorithms approach to this problem, such as solution representation, selection methods, genetic operators and repair of invalid solutions are presented. We also discuss an ACO for this problem. In ACO methodology, three strategies, ACO with only ranking, ACO with only max-min ant system (MMAS), and ACO with both ranking and MMAS, are considered. A comparison of these two approaches (i,e., GA and ACO) with the standard optimization method is presented to show the advantages of these approaches in terms of computation time.

  • PDF

A study on Improved Genetic Algorithm to solve nonlinear optimization problems (비선형 최적화문제의 해결을 위한 개선된 유전알고리즘의 연구)

  • 우병훈;하정진
    • Korean Management Science Review
    • /
    • v.13 no.1
    • /
    • pp.97-109
    • /
    • 1996
  • Genetic Algorithms have been successfully applied to various problems (for example, engineering design problems with a mix of continuous, integer and discrete design variables) that could not have been readily solved with traditional computational techniques. But, several problems for which conventional Genetic Algorithms are ill defined are premature convergence of solution and application of exterior penalty function. Therefore, we developed an Improved Genetic Algorithms (IGAs) to solve above two problems. As a case study, IGAs is applied to several nonlinear optimization problems and it is proved that this algorithm is very useful and efficient in comparison with traditional methods and conventional Genetic Algorithm.

  • PDF

Optimal Selection of Classifier Ensemble Using Genetic Algorithms (유전자 알고리즘을 이용한 분류자 앙상블의 최적 선택)

  • Kim, Myung-Jong
    • Journal of Intelligence and Information Systems
    • /
    • v.16 no.4
    • /
    • pp.99-112
    • /
    • 2010
  • Ensemble learning is a method for improving the performance of classification and prediction algorithms. It is a method for finding a highly accurateclassifier on the training set by constructing and combining an ensemble of weak classifiers, each of which needs only to be moderately accurate on the training set. Ensemble learning has received considerable attention from machine learning and artificial intelligence fields because of its remarkable performance improvement and flexible integration with the traditional learning algorithms such as decision tree (DT), neural networks (NN), and SVM, etc. In those researches, all of DT ensemble studies have demonstrated impressive improvements in the generalization behavior of DT, while NN and SVM ensemble studies have not shown remarkable performance as shown in DT ensembles. Recently, several works have reported that the performance of ensemble can be degraded where multiple classifiers of an ensemble are highly correlated with, and thereby result in multicollinearity problem, which leads to performance degradation of the ensemble. They have also proposed the differentiated learning strategies to cope with performance degradation problem. Hansen and Salamon (1990) insisted that it is necessary and sufficient for the performance enhancement of an ensemble that the ensemble should contain diverse classifiers. Breiman (1996) explored that ensemble learning can increase the performance of unstable learning algorithms, but does not show remarkable performance improvement on stable learning algorithms. Unstable learning algorithms such as decision tree learners are sensitive to the change of the training data, and thus small changes in the training data can yield large changes in the generated classifiers. Therefore, ensemble with unstable learning algorithms can guarantee some diversity among the classifiers. To the contrary, stable learning algorithms such as NN and SVM generate similar classifiers in spite of small changes of the training data, and thus the correlation among the resulting classifiers is very high. This high correlation results in multicollinearity problem, which leads to performance degradation of the ensemble. Kim,s work (2009) showedthe performance comparison in bankruptcy prediction on Korea firms using tradition prediction algorithms such as NN, DT, and SVM. It reports that stable learning algorithms such as NN and SVM have higher predictability than the unstable DT. Meanwhile, with respect to their ensemble learning, DT ensemble shows the more improved performance than NN and SVM ensemble. Further analysis with variance inflation factor (VIF) analysis empirically proves that performance degradation of ensemble is due to multicollinearity problem. It also proposes that optimization of ensemble is needed to cope with such a problem. This paper proposes a hybrid system for coverage optimization of NN ensemble (CO-NN) in order to improve the performance of NN ensemble. Coverage optimization is a technique of choosing a sub-ensemble from an original ensemble to guarantee the diversity of classifiers in coverage optimization process. CO-NN uses GA which has been widely used for various optimization problems to deal with the coverage optimization problem. The GA chromosomes for the coverage optimization are encoded into binary strings, each bit of which indicates individual classifier. The fitness function is defined as maximization of error reduction and a constraint of variance inflation factor (VIF), which is one of the generally used methods to measure multicollinearity, is added to insure the diversity of classifiers by removing high correlation among the classifiers. We use Microsoft Excel and the GAs software package called Evolver. Experiments on company failure prediction have shown that CO-NN is effectively applied in the stable performance enhancement of NNensembles through the choice of classifiers by considering the correlations of the ensemble. The classifiers which have the potential multicollinearity problem are removed by the coverage optimization process of CO-NN and thereby CO-NN has shown higher performance than a single NN classifier and NN ensemble at 1% significance level, and DT ensemble at 5% significance level. However, there remain further research issues. First, decision optimization process to find optimal combination function should be considered in further research. Secondly, various learning strategies to deal with data noise should be introduced in more advanced further researches in the future.

Developing Novel Algorithms to Reduce the Data Requirements of the Capture Matrix for a Wind Turbine Certification (풍력 발전기 평가를 위한 수집 행렬 데이터 절감 알고리즘 개발)

  • Lee, Jehyun;Choi, Jungchul
    • New & Renewable Energy
    • /
    • v.16 no.1
    • /
    • pp.15-24
    • /
    • 2020
  • For mechanical load testing of wind turbines, capture matrix is constructed for various range of wind speeds according to the international standard IEC 61400-13. The conventional method wastes considerable amount of data by its invalid data policy -segment data into 10 minutes then remove invalid ones. Previously, we have suggested an alternative way to save the total amount of data to build a capture matrix, but the efficient selection of data has been still under question. The paper introduces optimization algorithms to construct capture matrix with less data. Heuristic algorithm (simple stacking and lowest frequency first), population method (particle swarm optimization) and Q-Learning accompanied with epsilon-greedy exploration are compared. All algorithms show better performance than the conventional way, where the distribution of enhancement was quite diverse. Among the algorithms, the best performance was achieved by heuristic method (lowest frequency first), and similarly by particle swarm optimization: Approximately 28% of data reduction in average and more than 40% in maximum. On the other hand, unexpectedly, the worst performance was achieved by Q-Learning, which was a promising candidate at the beginning. This study is helpful for not only wind turbine evaluation particularly the viewpoint of cost, but also understanding nature of wind speed data.

Derivative Evaluation and Conditional Random Selection for Accelerating Genetic Algorithms

  • Jung, Sung-Hoon
    • International Journal of Fuzzy Logic and Intelligent Systems
    • /
    • v.5 no.1
    • /
    • pp.21-28
    • /
    • 2005
  • This paper proposes a new method for accelerating the search speed of genetic algorithms by taking derivative evaluation and conditional random selection into account in their evolution process. Derivative evaluation makes genetic algorithms focus on the individuals whose fitness is rapidly increased. This accelerates the search speed of genetic algorithms by enhancing exploitation like steepest descent methods but also increases the possibility of a premature convergence that means most individuals after a few generations approach to local optima. On the other hand, derivative evaluation under a premature convergence helps genetic algorithms escape the local optima by enhancing exploration. If GAs fall into a premature convergence, random selection is used in order to help escaping local optimum, but its effects are not large. We experimented our method with one combinatorial problem and five complex function optimization problems. Experimental results showed that our method was superior to the simple genetic algorithm especially when the search space is large.

MODIFIED LIMITED MEMORY BFGS METHOD WITH NONMONOTONE LINE SEARCH FOR UNCONSTRAINED OPTIMIZATION

  • Yuan, Gonglin;Wei, Zengxin;Wu, Yanlin
    • Journal of the Korean Mathematical Society
    • /
    • v.47 no.4
    • /
    • pp.767-788
    • /
    • 2010
  • In this paper, we propose two limited memory BFGS algorithms with a nonmonotone line search technique for unconstrained optimization problems. The global convergence of the given methods will be established under suitable conditions. Numerical results show that the presented algorithms are more competitive than the normal BFGS method.