• 제목/요약/키워드: multivariate optimization

검색결과 49건 처리시간 0.034초

하수처리장의 고도처리 upgrading 설계와 공정 최적화를 위한 다변량 통계분석 (Design of a Wastewater Treatment Plant Upgrading to Advanced Nutrient Removal Treatment Using Modeling Methodology and Multivariate Statistical Analysis for Process Optimization)

  • 김민정;김민한;김용수;유창규
    • Korean Chemical Engineering Research
    • /
    • 제48권5호
    • /
    • pp.589-597
    • /
    • 2010
  • 하수처리 시스템에서의 생물학적 영양염류 기준이 강화됨에 따라, 표준활성슬러지공법으로 운전 중인 하수처리장의 고도처리 공법으로의 개보수 필요성이 증가하고 있다. 그러나 실제 하수처리 시스템에서의 다양한 유입조건 및 운전조건의 복잡한 반응 구성으로 인해 실험을 통하여 개보수된 고도처리공법의 최적조건을 찾는 것은 쉽지 않은 일이며, 이는 많은 시간과 비용을 소모하여 비효율적이다. 따라서 본 연구에서는 활성슬러지공정모델(ASMs)을 기반으로 한 하수처리장의 모델링 및 시뮬레이션 기법을 통하여 하수처리장의 고도처리공법으로의 upgrading 설계를 수행하며, 이를 통계적이며 체계적으로 접근하기 위해 반응표면분석법(Response surface method)을 통한 고도처리공법의 설계 최적화를 수행하였다. 또한 실규모 하수처리장에서의 운전 최적화를 위해서는 하수처리의 동력학적 매개변수에 대한 정확한 분석이 수행되어야 한다. 본 연구에서는 다변량 통계분석 기법인 부분최소승자법(PLS)을 통하여 하수처리 시스템의 동력학적 매개변수 간의 상관관계를 파악하며, 고도처리공법 하수처리장의 운전 결과에 가장 큰 영향을 미치는 매개변수를 도출하였다. 본 연구를 통해 하수처리장의 고도처리공법 upgrading 설계 및 운전 최적화를 위한 방법론을 제시하였으며, 이를 통하여 설계시간 및 경비 절감 등 고도처리공법으로의 고효율적인 개보수가 가능할 것으로 예상된다.

Sampling Strategies for Computer Experiments: Design and Analysis

  • Lin, Dennis K.J.;Simpson, Timothy W.;Chen, Wei
    • International Journal of Reliability and Applications
    • /
    • 제2권3호
    • /
    • pp.209-240
    • /
    • 2001
  • Computer-based simulation and analysis is used extensively in engineering for a variety of tasks. Despite the steady and continuing growth of computing power and speed, the computational cost of complex high-fidelity engineering analyses and simulations limit their use in important areas like design optimization and reliability analysis. Statistical approximation techniques such as design of experiments and response surface methodology are becoming widely used in engineering to minimize the computational expense of running such computer analyses and circumvent many of these limitations. In this paper, we compare and contrast five experimental design types and four approximation model types in terms of their capability to generate accurate approximations for two engineering applications with typical engineering behaviors and a wide range of nonlinearity. The first example involves the analysis of a two-member frame that has three input variables and three responses of interest. The second example simulates the roll-over potential of a semi-tractor-trailer for different combinations of input variables and braking and steering levels. Detailed error analysis reveals that uniform designs provide good sampling for generating accurate approximations using different sample sizes while kriging models provide accurate approximations that are robust for use with a variety of experimental designs and sample sizes.

  • PDF

On the performance of improved quadrature spatial modulation

  • Holoubi, Tasnim;Murtala, Sheriff;Muchena, Nishal;Mohaisen, Manar
    • ETRI Journal
    • /
    • 제42권4호
    • /
    • pp.562-574
    • /
    • 2020
  • Quadrature spatial modulation (QSM) utilizes the in-phase and quadrature spatial dimensions to transmit the real and imaginary parts of a single signal symbol, respectively. The improved QSM (IQSM) transmits two signal symbols per channel use through a combination of two antennas for each of the real and imaginary parts. The main contributions of this study can be summarized as follows. First, we derive an upper bound for the error performance of the IQSM. We then design constellation sets that minimize the error performance of the IQSM for several system configurations. Second, we propose a double QSM (DQSM) that transmits the real and imaginary parts of two signal symbols through any available transmit antennas. Finally, we propose a parallel IQSM (PIQSM) that splits the antenna set into equal subsets and performs IQSM within each subset using the same two signal symbols. Simulation results demonstrate that the proposed constellations significantly outperform conventional constellations. Additionally, DQSM and PIQSM provide a performance similar to that of IQSM while requiring a smaller number of transmit antennas and outperform IQSM with the same number of transmit antennas.

개선된 데이터마이닝을 위한 혼합 학습구조의 제시 (Hybrid Learning Architectures for Advanced Data Mining:An Application to Binary Classification for Fraud Management)

  • Kim, Steven H.;Shin, Sung-Woo
    • 정보기술응용연구
    • /
    • 제1권
    • /
    • pp.173-211
    • /
    • 1999
  • The task of classification permeates all walks of life, from business and economics to science and public policy. In this context, nonlinear techniques from artificial intelligence have often proven to be more effective than the methods of classical statistics. The objective of knowledge discovery and data mining is to support decision making through the effective use of information. The automated approach to knowledge discovery is especially useful when dealing with large data sets or complex relationships. For many applications, automated software may find subtle patterns which escape the notice of manual analysis, or whose complexity exceeds the cognitive capabilities of humans. This paper explores the utility of a collaborative learning approach involving integrated models in the preprocessing and postprocessing stages. For instance, a genetic algorithm effects feature-weight optimization in a preprocessing module. Moreover, an inductive tree, artificial neural network (ANN), and k-nearest neighbor (kNN) techniques serve as postprocessing modules. More specifically, the postprocessors act as second0order classifiers which determine the best first-order classifier on a case-by-case basis. In addition to the second-order models, a voting scheme is investigated as a simple, but efficient, postprocessing model. The first-order models consist of statistical and machine learning models such as logistic regression (logit), multivariate discriminant analysis (MDA), ANN, and kNN. The genetic algorithm, inductive decision tree, and voting scheme act as kernel modules for collaborative learning. These ideas are explored against the background of a practical application relating to financial fraud management which exemplifies a binary classification problem.

  • PDF

Subset selection in multiple linear regression: An improved Tabu search

  • Bae, Jaegug;Kim, Jung-Tae;Kim, Jae-Hwan
    • Journal of Advanced Marine Engineering and Technology
    • /
    • 제40권2호
    • /
    • pp.138-145
    • /
    • 2016
  • This paper proposes an improved tabu search method for subset selection in multiple linear regression models. Variable selection is a vital combinatorial optimization problem in multivariate statistics. The selection of the optimal subset of variables is necessary in order to reliably construct a multiple linear regression model. Its applications widely range from machine learning, timeseries prediction, and multi-class classification to noise detection. Since this problem has NP-complete nature, it becomes more difficult to find the optimal solution as the number of variables increases. Two typical metaheuristic methods have been developed to tackle the problem: the tabu search algorithm and hybrid genetic and simulated annealing algorithm. However, these two methods have shortcomings. The tabu search method requires a large amount of computing time, and the hybrid algorithm produces a less accurate solution. To overcome the shortcomings of these methods, we propose an improved tabu search algorithm to reduce moves of the neighborhood and to adopt an effective move search strategy. To evaluate the performance of the proposed method, comparative studies are performed on small literature data sets and on large simulation data sets. Computational results show that the proposed method outperforms two metaheuristic methods in terms of the computing time and solution quality.

Response surface analysis of removal of a textile dye by a Turkish coal powder

  • Khataee, Alireza;Alidokht, Leila;Hassani, Aydin;Karaca, Semra
    • Advances in environmental research
    • /
    • 제2권4호
    • /
    • pp.291-308
    • /
    • 2013
  • In the present study, an experimental design methodology was used to optimize the adsorptive removal of Basic Yellow 13 (BY13) using Turkish coal powder. A central composite design (CCD) consisting of 31 experiments was employed to evaluate the simple and combined effects of the four independent variables, initial dye concentration (mg/L), adsorbent dosage (g/L), temperature ($^{\circ}C$) and contact time (min) on the color removal (CR) efficiency (%) and optimizing the process response. Analysis of variance (ANOVA) showed a high coefficient of determination value ($R^2=0.947$) and satisfactory prediction of the polynomial regression model was derived. Results indicated that the CR efficiency was not significantly affected by temperature in the range of $12-60^{\circ}C$. While all other variables significantly influenced response. The highest CR (95.14%), estimated by multivariate experimental design, was found at the optimal experimental conditions of initial dye concentration 30 mg/L, adsorbent dosage 1.5 g/L, temperature $25^{\circ}C$ and contact time 10 min.

A Comparative Study of Estimation by Analogy using Data Mining Techniques

  • Nagpal, Geeta;Uddin, Moin;Kaur, Arvinder
    • Journal of Information Processing Systems
    • /
    • 제8권4호
    • /
    • pp.621-652
    • /
    • 2012
  • Software Estimations provide an inclusive set of directives for software project developers, project managers, and the management in order to produce more realistic estimates based on deficient, uncertain, and noisy data. A range of estimation models are being explored in the industry, as well as in academia, for research purposes but choosing the best model is quite intricate. Estimation by Analogy (EbA) is a form of case based reasoning, which uses fuzzy logic, grey system theory or machine-learning techniques, etc. for optimization. This research compares the estimation accuracy of some conventional data mining models with a hybrid model. Different data mining models are under consideration, including linear regression models like the ordinary least square and ridge regression, and nonlinear models like neural networks, support vector machines, and multivariate adaptive regression splines, etc. A precise and comprehensible predictive model based on the integration of GRA and regression has been introduced and compared. Empirical results have shown that regression when used with GRA gives outstanding results; indicating that the methodology has great potential and can be used as a candidate approach for software effort estimation.

Improvement of inspection system for common crossings by track side monitoring and prognostics

  • Sysyn, Mykola;Nabochenko, Olga;Kovalchuk, Vitalii;Gruen, Dimitri;Pentsak, Andriy
    • Structural Monitoring and Maintenance
    • /
    • 제6권3호
    • /
    • pp.219-235
    • /
    • 2019
  • Scheduled inspections of common crossings are one of the main cost drivers of railway maintenance. Prognostics and health management (PHM) approach and modern monitoring means offer many possibilities in the optimization of inspections and maintenance. The present paper deals with data driven prognosis of the common crossing remaining useful life (RUL) that is based on an inertial monitoring system. The problem of scheduled inspections system for common crossings is outlined and analysed. The proposed analysis of inertial signals with the maximal overlap discrete wavelet packet transform (MODWPT) and Shannon entropy (SE) estimates enable to extract the spectral features. The relevant features for the acceleration components are selected with application of Lasso (Least absolute shrinkage and selection operator) regularization. The features are fused with time domain information about the longitudinal position of wheels impact and train velocities by multivariate regression. The fused structural health (SH) indicator has a significant correlation to the lifetime of crossing. The RUL prognosis is performed on the linear degradation stochastic model with recursive Bayesian update. Prognosis testing metrics show the promising results for common crossing inspection scheduling improvement.

한국 프로스포츠 선수들의 연봉에 대한 다변량적 분석 (A Multivariate Analysis of Korean Professional Players Salary)

  • 송종우
    • 응용통계연구
    • /
    • 제21권3호
    • /
    • pp.441-453
    • /
    • 2008
  • 프로스포츠 선수들의 연봉은 선수들의 개인 성적과 팀에 대한 기여도 등으로 결정된다는 가정하에 프로농구와 프로야구 선수들의 전년도 성적으로 다음해 연봉을 예측 분석하였다. 분석에 있어서 data visualization 기법을 통해 변수사이의 관계, 이상점 발견, 모형진단등을 하였다. 다중선형회귀 모형(Multiple Linear Regression)과 트리모형(Regression Tree)을 이용해서 자료를 분석하고 모델간 비교를 했으며, Cross-Validation을 이용해서 최적모델을 선택하였다. 특히, 자동으로 변수선택을 하는 stepwise regression방법을 그냥 사용하기보다는 먼저 설명변수들 사이의 관계나 설명변수와 반응변수 사이의 관계등을 조사하고 나서 이를 통해 선택된 변수들을 가지고 stepwise regression과 regression tree 방법론을 이용해서 적절한 변수 및 최종 모형을 선택하였다. 분석결과, 프로농구의 경우에는 경기당 득점, 어시스트, 자유투 성공수, 경력 등이 중요한 변수였고, 프로야구 투수의 경우에는 경력, 9이닝 당 삼진 수, 방어율, 피홈런 수 등이 중요한 변수였고, 프로야구 타자의 경우에는 경력, 안타 수, FA(자유계약)유무 여부 등이 중요한 변수였다.

비정렬 다변수 데이터의 B-스플라인 근사화 기법 (On B-spline Approximation for Representing Scattered Multivariate Data)

  • 박상근
    • 대한기계학회논문집A
    • /
    • 제35권8호
    • /
    • pp.921-931
    • /
    • 2011
  • 본 연구는 B-스플라인 하이퍼볼륨을 사용하여 주어진 비정렬 데이터를 근사화하는 데이터 근사기법에 관한 것이다. 개발 구현을 위한 B-스플라인 하이퍼볼륨의 자료 구조가 기술되며 해당 메모리 크기의 측정을 통해 간결한 표현 모델임을 보인다. 제안하는 근사 기법은 두 가지 알고리즘으로 구성된다. 하나는 B-스플라인 하이퍼볼륨의 절점 벡터 결정에 관한 것이고, 다른 하나는 조정점 결정에 관한 것으로 최소자승 최소화 문제의 해를 구함으로써 얻게 된다. 여기서 구한 해는 데이터 복잡성에 의존하지 않는다. 본 연구 방식은 다양한 형태의 데이터 분포를 가지고 근사 정밀도, 메모리 사용량, 계산 시간 등의 근사화 성능(수준)을 평가한다. 더불어 기존 방법과의 비교를 통해 유용성을 보이며, 비구속 최적화 예제를 통하여 다양한 응용 분야로의 가능성을 보여준다.