• Title/Summary/Keyword: Sampling-Based Algorithm

Search Result 477, Processing Time 0.026 seconds

Optimization of a horizontal axis marine current turbine via surrogate models

  • Thandayutham, Karthikeyan;Avital, E.J.;Venkatesan, Nithya;Samad, Abdus
    • Ocean Systems Engineering
    • /
    • v.9 no.2
    • /
    • pp.111-133
    • /
    • 2019
  • Flow through a scaled horizontal axis marine current turbine was numerically simulated after validation and the turbine design was optimized. The computational fluid dynamics (CFD) code Ansys-CFX 16.1 for numerical modeling, an in-house blade element momentum (BEM) code for analytical modeling and an in-house surrogate-based optimization (SBO) code were used to find an optimal turbine design. The blade-pitch angle (${\theta}$) and the number of rotor blades (NR) were taken as design variables. A single objective optimization approach was utilized in the present work. The defined objective function was the turbine's power coefficient ($C_P$). A $3{\times}3$ full-factorial sampling technique was used to define the sample space. This sampling technique gave different turbine designs, which were further evaluated for the objective function by solving the Reynolds-Averaged Navier-Stokes equations (RANS). Finally, the SBO technique with search algorithm produced an optimal design. It is found that the optimal design has improved the objective function by 26.5%. This article presents the solution approach, analysis of the turbine flow field and the predictability of various surrogate based techniques.

Bayesian estimation of kinematic parameters of disk galaxies in large HI galaxy surveys

  • Oh, Se-Heon;Staveley-Smith, Lister
    • The Bulletin of The Korean Astronomical Society
    • /
    • v.41 no.2
    • /
    • pp.62.2-62.2
    • /
    • 2016
  • We present a newly developed algorithm based on a Bayesian method for 2D tilted-ring analysis of disk galaxies which operates on velocity fields. Compared to the conventional ones based on a chi-squared minimisation procedure, this new Bayesian-based algorithm less suffers from local minima of the model parameters even with high multi-modality of their posterior distributions. Moreover, the Bayesian analysis implemented via Markov Chain Monte Carlo (MCMC) sampling only requires broad ranges of posterior distributions of the parameters, which makes the fitting procedure fully automated. This feature is essential for performing kinematic analysis of an unprecedented number of resolved galaxies from the upcoming Square Kilometre Array (SKA) pathfinders' galaxy surveys. A standalone code, the so-called '2D Bayesian Automated Tilted-ring fitter' (2DBAT) that implements the Bayesian fits of 2D tilted-ring models is developed for deriving rotation curves of galaxies that are at least marginally resolved (> 3 beams across the semi-major axis) and moderately inclined (20 < i < 70 degree). The main layout of 2DBAT and its performance test are discussed using sample galaxies from Australia Telescope Compact Array (ATCA) observations as well as artificial data cubes built based on representative rotation curves of intermediate-mass and massive spiral galaxies.

  • PDF

Imbalanced SVM-Based Anomaly Detection Algorithm for Imbalanced Training Datasets

  • Wang, GuiPing;Yang, JianXi;Li, Ren
    • ETRI Journal
    • /
    • v.39 no.5
    • /
    • pp.621-631
    • /
    • 2017
  • Abnormal samples are usually difficult to obtain in production systems, resulting in imbalanced training sample sets. Namely, the number of positive samples is far less than the number of negative samples. Traditional Support Vector Machine (SVM)-based anomaly detection algorithms perform poorly for highly imbalanced datasets: the learned classification hyperplane skews toward the positive samples, resulting in a high false-negative rate. This article proposes a new imbalanced SVM (termed ImSVM)-based anomaly detection algorithm, which assigns a different weight for each positive support vector in the decision function. ImSVM adjusts the learned classification hyperplane to make the decision function achieve a maximum GMean measure value on the dataset. The above problem is converted into an unconstrained optimization problem to search the optimal weight vector. Experiments are carried out on both Cloud datasets and Knowledge Discovery and Data Mining datasets to evaluate ImSVM. Highly imbalanced training sample sets are constructed. The experimental results show that ImSVM outperforms over-sampling techniques and several existing imbalanced SVM-based techniques.

Automation of Sampling for Public Survey Performance Assessment (공공측량 성과심사 표본추출 자동화 가능성 분석)

  • Choi, Hyun;Jin, Cheol;Lee, Jung Il;Kim, Gi Hong
    • KSCE Journal of Civil and Environmental Engineering Research
    • /
    • v.44 no.1
    • /
    • pp.95-100
    • /
    • 2024
  • The public survey performance review conducted by the Spatial Information Quality Management Institute is conducted at the screening rate in accordance with the regulations, and the examiner directly judges the overall trend of the submitted performance based on the extracted sample. However, the evaluation of the Ministry of Land, Infrastructure and Transport, the evaluation trustee shall be specified by random extraction (Random Collection) is specified by the sample. In this study, it analyzed the details of the actual site and analyzed through securing actual performance review data. In addition, we analyzed considerations according to various field conditions and studied ways to apply the public survey performance review sampling algorithm. Therefore, detailed sampling criteria analysis by performance reviewers is necessary. A relative comparison was made feasible by comparing the data for which the real performance evaluation was performed with the outcomes of the Python automation program. This automation program is expected to be employed as a foundation program for the automated application of public survey performance evaluation sampling in the future.

An Efficient Method to Compute a Covariance Matrix of the Non-local Means Algorithm for Image Denoising with the Principal Component Analysis (영상 잡음 제거를 위한 주성분 분석 기반 비 지역적 평균 알고리즘의 효율적인 공분산 행렬 계산 방법)

  • Kim, Jeonghwan;Jeong, Jechang
    • Journal of Broadcast Engineering
    • /
    • v.21 no.1
    • /
    • pp.60-65
    • /
    • 2016
  • This paper introduces the non-local means (NLM) algorithm for image denoising, and also introduces an improved algorithm which is based on the principal component analysis (PCA). To do the PCA, a covariance matrix of a given image should be evaluated first. If we let the size of neighborhood patches of the NLM S × S2, and let the number of pixels Q, a matrix multiplication of the size S2 × Q is required to compute a covariance matrix. According to the characteristic of images, such computation is inefficient. Therefore, this paper proposes an efficient method to compute the covariance matrix by sampling the pixels. After sampling, the covariance matrix can be computed with matrices of the size S2 × floor (Width/l) × (Height/l).

Parallelization of Probabilistic RoadMap for Generating UAV Path on a DTED Map (DTED 맵에서 무인기 경로 생성을 위한 Probabilistic RoadMap 병렬화)

  • Noh, Geemoon;Park, Jihoon;Min, Chanoh;Lee, Daewoo
    • Journal of the Korean Society for Aeronautical & Space Sciences
    • /
    • v.50 no.3
    • /
    • pp.157-164
    • /
    • 2022
  • In this paper, we describe how to implement the mountainous terrain, radar, and air defense network for UAV path planning in a 3-D environment, and perform path planning and re-planning using the PRM algorithm, a sampling-based path planning algorithm. In the case of the original PRM algorithm, the calculation to check whether there is an obstacle between the nodes is performed 1:1 between nodes and is performed continuously, so the amount of calculation is greatly affected by the number of nodes or the linked distance between nodes. To improve this part, the proposed LineGridMask method simplifies the method of checking whether obstacles exist, and reduces the calculation time of the path planning through parallelization. Finally, comparing performance with existing PRM algorithms confirmed that computational time was reduced by up to 88% in path planning and up to 94% in re-planning.

Analysis of Performance for Entropy-Based ISAR Autofocus Technique (엔트로피 기반의 ISAR 자동 초점 기법에 대한 성능 분석)

  • Bae, Jun-Woo;Kim, Kyung-Tae;Lee, Jin-Ho;Im, Jeong-Heom
    • The Journal of Korean Institute of Electromagnetic Engineering and Science
    • /
    • v.17 no.12 s.115
    • /
    • pp.1249-1258
    • /
    • 2006
  • Two-dimensional(2-D) radar images, namely, ISAR images from a maneuvering target include unwanted phase errors due to the target's motion. These phase errors make ISAR images to be blurred. The ISAR autofocus technique is required in order to remove these unwanted phase errors. Unless those unwanted phase errors produced by the target's motion are removed prior to target identification, we cannot expect a reliable target identification performance. In this paper, we use the entropy-based ISAR autofocus technique which consists of two steps: range alignment and phase adjustment. We analyze a relationship between the number of sampling point and a image quality in a range alignment algorithm and also analyze a technique for reducing computation time of the SSA(Stage-by-Stage Approachng) algorithm in a phase adjustment.

Response Modeling for the Marketing Promotion with Weighted Case Based Reasoning Under Imbalanced Data Distribution (불균형 데이터 환경에서 변수가중치를 적용한 사례기반추론 기반의 고객반응 예측)

  • Kim, Eunmi;Hong, Taeho
    • Journal of Intelligence and Information Systems
    • /
    • v.21 no.1
    • /
    • pp.29-45
    • /
    • 2015
  • Response modeling is a well-known research issue for those who have tried to get more superior performance in the capability of predicting the customers' response for the marketing promotion. The response model for customers would reduce the marketing cost by identifying prospective customers from very large customer database and predicting the purchasing intention of the selected customers while the promotion which is derived from an undifferentiated marketing strategy results in unnecessary cost. In addition, the big data environment has accelerated developing the response model with data mining techniques such as CBR, neural networks and support vector machines. And CBR is one of the most major tools in business because it is known as simple and robust to apply to the response model. However, CBR is an attractive data mining technique for data mining applications in business even though it hasn't shown high performance compared to other machine learning techniques. Thus many studies have tried to improve CBR and utilized in business data mining with the enhanced algorithms or the support of other techniques such as genetic algorithm, decision tree and AHP (Analytic Process Hierarchy). Ahn and Kim(2008) utilized logit, neural networks, CBR to predict that which customers would purchase the items promoted by marketing department and tried to optimized the number of k for k-nearest neighbor with genetic algorithm for the purpose of improving the performance of the integrated model. Hong and Park(2009) noted that the integrated approach with CBR for logit, neural networks, and Support Vector Machine (SVM) showed more improved prediction ability for response of customers to marketing promotion than each data mining models such as logit, neural networks, and SVM. This paper presented an approach to predict customers' response of marketing promotion with Case Based Reasoning. The proposed model was developed by applying different weights to each feature. We deployed logit model with a database including the promotion and the purchasing data of bath soap. After that, the coefficients were used to give different weights of CBR. We analyzed the performance of proposed weighted CBR based model compared to neural networks and pure CBR based model empirically and found that the proposed weighted CBR based model showed more superior performance than pure CBR model. Imbalanced data is a common problem to build data mining model to classify a class with real data such as bankruptcy prediction, intrusion detection, fraud detection, churn management, and response modeling. Imbalanced data means that the number of instance in one class is remarkably small or large compared to the number of instance in other classes. The classification model such as response modeling has a lot of trouble to recognize the pattern from data through learning because the model tends to ignore a small number of classes while classifying a large number of classes correctly. To resolve the problem caused from imbalanced data distribution, sampling method is one of the most representative approach. The sampling method could be categorized to under sampling and over sampling. However, CBR is not sensitive to data distribution because it doesn't learn from data unlike machine learning algorithm. In this study, we investigated the robustness of our proposed model while changing the ratio of response customers and nonresponse customers to the promotion program because the response customers for the suggested promotion is always a small part of nonresponse customers in the real world. We simulated the proposed model 100 times to validate the robustness with different ratio of response customers to response customers under the imbalanced data distribution. Finally, we found that our proposed CBR based model showed superior performance than compared models under the imbalanced data sets. Our study is expected to improve the performance of response model for the promotion program with CBR under imbalanced data distribution in the real world.

Super-resolution Algorithm using Discrete Wavelet Transform for Single-image (이산 웨이블릿 변환을 이용한 영상의 초고해상도 기법)

  • Lim, Jong-Myeong;Yoo, Ji-Sang
    • Journal of Broadcast Engineering
    • /
    • v.17 no.2
    • /
    • pp.344-353
    • /
    • 2012
  • In this paper, we propose a super-resolution algorithm using discrete wavelet transform. In general super-resolution algorithms for single-image, probability based operations have been used for searching high-frequency components. Consequently, the complexity of the algorithm causes the increase of processing time. In the proposed algorithm, we use discrete wavelet transform to find high-frequency sub-bands. We perform inverse discrete wavelet transform using input image and high-frequency sub-bands of the same resolution as the input image which are obtained by performing discrete wavelet transform without down-sampling and then we obtain image with high-resolution. In the proposed algorithm, we use the down-sampled version of the original image ($512{\times}512$) as a test image ($256{\times}256$) to compare the performance of algorithms. Through experimental results, we confirm the improved efficiency of the proposed algorithm comparing with conventional interpolation algorithms and also decreased processing time comparing the probability based operations.

Optimization of water quality monitoring stations using genetic algorithm, a case study, Sefid-Rud River, Iran

  • Asadollahfardi, Gholamreza;Heidarzadeh, Nima;Mosalli, Atabak;Sekhavati, Ali
    • Advances in environmental research
    • /
    • v.7 no.2
    • /
    • pp.87-107
    • /
    • 2018
  • Water quality monitoring network needs periodic evaluations based on environmental demands and financial constraints. We used a genetic algorithm to optimize the existing water quality monitoring stations on the Sefid-Rud River, which is located in the North of Iran. Our objective was to optimize the existing stations for drinking and irrigation purposes, separately. The technique includes two stages called data preparation and the optimization. On the data preparation stage, first the basin was divided into four sections and each section was consisted of some stations. Then, the score of each station was computed using the data provided by the water Research Institute of the Ministry of energy. After that, we applied a weighting method by providing questionnaires to ask the experts to define the significance of each parameter. In the next step, according to the scores, stations were prioritized cumulatively. Finally, the genetic algorithm was applied to identify the best combination. The results indicated that out of 21 existing monitoring stations, 14 stations should remain in the network for both irrigation and drinking purposes. The results also had a good compliance with the previous studies which used dynamic programming as the optimization technique.