• Title/Summary/Keyword: Discrete Optimization

Search Result 508, Processing Time 0.031 seconds

Optimum Design of Pitch Reducer for Wind Turbine Using Genetic Algorithm (유전 알고리즘을 이용한 풍력발전기용 피치감속기의 최적 설계)

  • Kim, Jeong Gil;Park, Young Jun;Lee, Geun Ho;Nam, Yong Yun;Yang, Woo Yeoul
    • Transactions of the Korean Society of Mechanical Engineers A
    • /
    • v.38 no.2
    • /
    • pp.185-192
    • /
    • 2014
  • Planetary gear design is complex because it involves a combination of discrete variables such as module, integer variables such as the number of teeth, and continuous variables such as face width and aspect ratio. Thus, an optimum design technique is needed. In this study, we applied a genetic algorithm to the design optimization of a planetary gear. In this algorithm, tooth root strength and surface durability are assessed with fundamental variables such as the number of teeth, module, pressure angle, and face width. With the help of this technique, gear designers could reduce trial and error in the initial design stages, thus cutting the time required for planetary gear design.

Complexity Estimation Based Work Load Balancing for a Parallel Lidar Waveform Decomposition Algorithm

  • Jung, Jin-Ha;Crawford, Melba M.;Lee, Sang-Hoon
    • Korean Journal of Remote Sensing
    • /
    • v.25 no.6
    • /
    • pp.547-557
    • /
    • 2009
  • LIDAR (LIght Detection And Ranging) is an active remote sensing technology which provides 3D coordinates of the Earth's surface by performing range measurements from the sensor. Early small footprint LIDAR systems recorded multiple discrete returns from the back-scattered energy. Recent advances in LIDAR hardware now make it possible to record full digital waveforms of the returned energy. LIDAR waveform decomposition involves separating the return waveform into a mixture of components which are then used to characterize the original data. The most common statistical mixture model used for this process is the Gaussian mixture. Waveform decomposition plays an important role in LIDAR waveform processing, since the resulting components are expected to represent reflection surfaces within waveform footprints. Hence the decomposition results ultimately affect the interpretation of LIDAR waveform data. Computational requirements in the waveform decomposition process result from two factors; (1) estimation of the number of components in a mixture and the resulting parameter estimates, which are inter-related and cannot be solved separately, and (2) parameter optimization does not have a closed form solution, and thus needs to be solved iteratively. The current state-of-the-art airborne LIDAR system acquires more than 50,000 waveforms per second, so decomposing the enormous number of waveforms is challenging using traditional single processor architecture. To tackle this issue, four parallel LIDAR waveform decomposition algorithms with different work load balancing schemes - (1) no weighting, (2) a decomposition results-based linear weighting, (3) a decomposition results-based squared weighting, and (4) a decomposition time-based linear weighting - were developed and tested with varying number of processors (8-256). The results were compared in terms of efficiency. Overall, the decomposition time-based linear weighting work load balancing approach yielded the best performance among four approaches.

A Finite Memory Structure Smoothing Filter and Its Equivalent Relationship with Existing Filters (유한기억구조 스무딩 필터와 기존 필터와의 등가 관계)

  • Kim, Min Hui;Kim, Pyung Soo
    • KIPS Transactions on Computer and Communication Systems
    • /
    • v.10 no.2
    • /
    • pp.53-58
    • /
    • 2021
  • In this paper, an alternative finite memory structure(FMS) smoothing filter is developed for discrete-time state-space model with a control input. To obtain the FMS smoothing filter, unbiasedness will be required beforehand in addition to a performance criteria of minimum variance. The FMS smoothing filter is obtained by directly solving an optimization problem with the unbiasedness constraint using only finite measurements and inputs on the most recent window. The proposed FMS smoothing filter is shown to have intrinsic good properties such as deadbeat and time-invariance. In addition, the proposed FMS smoothing filter is shown to be equivalent to existing FMS filters according to the delay length between the measurement and the availability of its estimate. Finally, to verify intrinsic robustness of the proposed FMS smoothing filter, computer simulations are performed for a temporary model uncertainty. Simulation results show that the proposed FMS smoothing filter can be better than the standard FMS filter and Kalman filter.

Aircraft application with artificial fuzzy heuristic theory via drone

  • C.C. Hung;T. Nguyen;C.Y. Hsieh
    • Advances in aircraft and spacecraft science
    • /
    • v.10 no.6
    • /
    • pp.495-519
    • /
    • 2023
  • The drone serves the customers not served by vans. At the same time, considering the safety, policy and terrain as well as the need to replace the battery, the drone needs to be transported by truck to the identified station along with the parcel. From each such station, the drone serves a subset of customers according to a direct assignment pattern, i.e., every time the drone is launched, it serves one demand node and returns to the station to collect another parcel. Similarly, the truck is used to transport the drone and cargo between stations. This is somewhat different from the research of other scholars. In terms of the joint distribution of the drone and road vehicle, most scholars will choose the combination of two transportation tools, while we use three. The drone and vans are responsible for distribution services, and the trucks are responsible for transporting the goods and drone to the station. The goal is to optimize the total delivery cost which includes the transportation costs for the vans and the delivery cost for the drone. A fixed cost is also considered for each drone parking site corresponding to the cost of positioning the drone and using the drone station. A discrete optimization model is presented for the problem in addition to a two-phase heuristic algorithm. The results of a series of computational tests performed to assess the applicability of the model and the efficiency of the heuristic are reported. The results obtained show that nearly 10% of the cost can be saved by combining the traditional delivery mode with the use of a drone and drone stations.

Wavelet Thresholding Techniques to Support Multi-Scale Decomposition for Financial Forecasting Systems

  • Shin, Taeksoo;Han, Ingoo
    • Proceedings of the Korea Database Society Conference
    • /
    • 1999.06a
    • /
    • pp.175-186
    • /
    • 1999
  • Detecting the features of significant patterns from their own historical data is so much crucial to good performance specially in time-series forecasting. Recently, a new data filtering method (or multi-scale decomposition) such as wavelet analysis is considered more useful for handling the time-series that contain strong quasi-cyclical components than other methods. The reason is that wavelet analysis theoretically makes much better local information according to different time intervals from the filtered data. Wavelets can process information effectively at different scales. This implies inherent support fer multiresolution analysis, which correlates with time series that exhibit self-similar behavior across different time scales. The specific local properties of wavelets can for example be particularly useful to describe signals with sharp spiky, discontinuous or fractal structure in financial markets based on chaos theory and also allows the removal of noise-dependent high frequencies, while conserving the signal bearing high frequency terms of the signal. To date, the existing studies related to wavelet analysis are increasingly being applied to many different fields. In this study, we focus on several wavelet thresholding criteria or techniques to support multi-signal decomposition methods for financial time series forecasting and apply to forecast Korean Won / U.S. Dollar currency market as a case study. One of the most important problems that has to be solved with the application of the filtering is the correct choice of the filter types and the filter parameters. If the threshold is too small or too large then the wavelet shrinkage estimator will tend to overfit or underfit the data. It is often selected arbitrarily or by adopting a certain theoretical or statistical criteria. Recently, new and versatile techniques have been introduced related to that problem. Our study is to analyze thresholding or filtering methods based on wavelet analysis that use multi-signal decomposition algorithms within the neural network architectures specially in complex financial markets. Secondly, through the comparison with different filtering techniques' results we introduce the present different filtering criteria of wavelet analysis to support the neural network learning optimization and analyze the critical issues related to the optimal filter design problems in wavelet analysis. That is, those issues include finding the optimal filter parameter to extract significant input features for the forecasting model. Finally, from existing theory or experimental viewpoint concerning the criteria of wavelets thresholding parameters we propose the design of the optimal wavelet for representing a given signal useful in forecasting models, specially a well known neural network models.

  • PDF

Wavelet Thresholding Techniques to Support Multi-Scale Decomposition for Financial Forecasting Systems

  • Shin, Taek-Soo;Han, In-Goo
    • Proceedings of the Korea Inteligent Information System Society Conference
    • /
    • 1999.03a
    • /
    • pp.175-186
    • /
    • 1999
  • Detecting the features of significant patterns from their own historical data is so much crucial to good performance specially in time-series forecasting. Recently, a new data filtering method (or multi-scale decomposition) such as wavelet analysis is considered more useful for handling the time-series that contain strong quasi-cyclical components than other methods. The reason is that wavelet analysis theoretically makes much better local information according to different time intervals from the filtered data. Wavelets can process information effectively at different scales. This implies inherent support for multiresolution analysis, which correlates with time series that exhibit self-similar behavior across different time scales. The specific local properties of wavelets can for example be particularly useful to describe signals with sharp spiky, discontinuous or fractal structure in financial markets based on chaos theory and also allows the removal of noise-dependent high frequencies, while conserving the signal bearing high frequency terms of the signal. To data, the existing studies related to wavelet analysis are increasingly being applied to many different fields. In this study, we focus on several wavelet thresholding criteria or techniques to support multi-signal decomposition methods for financial time series forecasting and apply to forecast Korean Won / U.S. Dollar currency market as a case study. One of the most important problems that has to be solved with the application of the filtering is the correct choice of the filter types and the filter parameters. If the threshold is too small or too large then the wavelet shrinkage estimator will tend to overfit or underfit the data. It is often selected arbitrarily or by adopting a certain theoretical or statistical criteria. Recently, new and versatile techniques have been introduced related to that problem. Our study is to analyze thresholding or filtering methods based on wavelet analysis that use multi-signal decomposition algorithms within the neural network architectures specially in complex financial markets. Secondly, through the comparison with different filtering techniques results we introduce the present different filtering criteria of wavelet analysis to support the neural network learning optimization and analyze the critical issues related to the optimal filter design problems in wavelet analysis. That is, those issues include finding the optimal filter parameter to extract significant input features for the forecasting model. Finally, from existing theory or experimental viewpoint concerning the criteria of wavelets thresholding parameters we propose the design of the optimal wavelet for representing a given signal useful in forecasting models, specially a well known neural network models.

  • PDF

Principles of Simulated Moving Bed Reactor(SMBR) (Simulated Moving Bed Reactor(SMBR)의 원리)

  • Song, Jae-Ryong;Kim, Jin-Il;Koo, Yoon-Mo
    • Korean Chemical Engineering Research
    • /
    • v.49 no.2
    • /
    • pp.129-136
    • /
    • 2011
  • Simulated Moving Bed(SMB) process consists of multiple chromatographic columns, which are usually partitioned into four zones. Such a process characteristic allows a continuous binary separations those are impracticable in conventional batch chromatographic processes. Compared with batch chromatography, SMB has advantages of continuity, high purity and productivity. Various researches have been reported for the integration of reaction and recovery during process operation on the purpose of economics and effectiveness. Simulated Moving Bed Reactor(SMBR) is introduced to combine SMB as a continuous separation process and reactor. Several cases of SMBR have been reported for diverse reactions with catalytic, enzymatic and chemical reaction on ion exchange resin as main streams. With an early type of fixed bed using catalyst, SMBR has been developed as SMB using fluidized enzyme, SMB with immobilized enzyme and SMB with discrete reaction region. For simple modeling and optimization of SMBR, a method considering convection only is possible. A complex method considering axial dispersion and mass transfer resistance is needed to explain the real behavior of solutes in SMBR. By combining reaction and separation, SMBR has benefits of lower installation cost by minimizing equipment use, higher purity and yield by avoiding the equilibrium restriction in case of reversible reaction.

An Intelligent Intrusion Detection Model Based on Support Vector Machines and the Classification Threshold Optimization for Considering the Asymmetric Error Cost (비대칭 오류비용을 고려한 분류기준값 최적화와 SVM에 기반한 지능형 침입탐지모형)

  • Lee, Hyeon-Uk;Ahn, Hyun-Chul
    • Journal of Intelligence and Information Systems
    • /
    • v.17 no.4
    • /
    • pp.157-173
    • /
    • 2011
  • As the Internet use explodes recently, the malicious attacks and hacking for a system connected to network occur frequently. This means the fatal damage can be caused by these intrusions in the government agency, public office, and company operating various systems. For such reasons, there are growing interests and demand about the intrusion detection systems (IDS)-the security systems for detecting, identifying and responding to unauthorized or abnormal activities appropriately. The intrusion detection models that have been applied in conventional IDS are generally designed by modeling the experts' implicit knowledge on the network intrusions or the hackers' abnormal behaviors. These kinds of intrusion detection models perform well under the normal situations. However, they show poor performance when they meet a new or unknown pattern of the network attacks. For this reason, several recent studies try to adopt various artificial intelligence techniques, which can proactively respond to the unknown threats. Especially, artificial neural networks (ANNs) have popularly been applied in the prior studies because of its superior prediction accuracy. However, ANNs have some intrinsic limitations such as the risk of overfitting, the requirement of the large sample size, and the lack of understanding the prediction process (i.e. black box theory). As a result, the most recent studies on IDS have started to adopt support vector machine (SVM), the classification technique that is more stable and powerful compared to ANNs. SVM is known as a relatively high predictive power and generalization capability. Under this background, this study proposes a novel intelligent intrusion detection model that uses SVM as the classification model in order to improve the predictive ability of IDS. Also, our model is designed to consider the asymmetric error cost by optimizing the classification threshold. Generally, there are two common forms of errors in intrusion detection. The first error type is the False-Positive Error (FPE). In the case of FPE, the wrong judgment on it may result in the unnecessary fixation. The second error type is the False-Negative Error (FNE) that mainly misjudges the malware of the program as normal. Compared to FPE, FNE is more fatal. Thus, when considering total cost of misclassification in IDS, it is more reasonable to assign heavier weights on FNE rather than FPE. Therefore, we designed our proposed intrusion detection model to optimize the classification threshold in order to minimize the total misclassification cost. In this case, conventional SVM cannot be applied because it is designed to generate discrete output (i.e. a class). To resolve this problem, we used the revised SVM technique proposed by Platt(2000), which is able to generate the probability estimate. To validate the practical applicability of our model, we applied it to the real-world dataset for network intrusion detection. The experimental dataset was collected from the IDS sensor of an official institution in Korea from January to June 2010. We collected 15,000 log data in total, and selected 1,000 samples from them by using random sampling method. In addition, the SVM model was compared with the logistic regression (LOGIT), decision trees (DT), and ANN to confirm the superiority of the proposed model. LOGIT and DT was experimented using PASW Statistics v18.0, and ANN was experimented using Neuroshell 4.0. For SVM, LIBSVM v2.90-a freeware for training SVM classifier-was used. Empirical results showed that our proposed model based on SVM outperformed all the other comparative models in detecting network intrusions from the accuracy perspective. They also showed that our model reduced the total misclassification cost compared to the ANN-based intrusion detection model. As a result, it is expected that the intrusion detection model proposed in this paper would not only enhance the performance of IDS, but also lead to better management of FNE.