• Title/Summary/Keyword: Fuzzy Linear Regression

Search Result 118, Processing Time 0.023 seconds

Robust Parameter Estimation using Fuzzy RANSAC (퍼지 RANSAC을 이용한 강건한 인수 예측)

  • Lee Joong-Jae;Jang Hyo-Jong;Kim Gye-Young;Choi Hyung-il
    • Journal of KIISE:Software and Applications
    • /
    • v.33 no.2
    • /
    • pp.252-266
    • /
    • 2006
  • Many problems in computer vision are mainly based on mathematical models. Their optimal solutions can be found by estimating the parameters of each model. However, provided an input data set is involved outliers which are relative]V larger than normal noises, they lead to incorrect results. RANSAC is a representative robust algorithm which is used to resolve the problem. One major problem with RANSAC is that it needs priori knowledge(i.e. a percentage of outliers) of the distribution of data. To solve this problem, we propose a FRANSAC algorithm which improves the rejection rate of outliers and the accuracy of solutions. This is peformed by categorizing all data into good sample set, bad sample set and vague sample set using a fuzzy classification at each iteration and sampling in only good sample set. In the experimental results, we show that the performance of the proposed algorithm when it is applied to the linear regression and the calculation of a homography.

A Hybrid QFD Framework for New Product Development

  • Tsai, Y-C;Chin, K-S;Yang, J-B
    • International Journal of Quality Innovation
    • /
    • v.3 no.2
    • /
    • pp.138-158
    • /
    • 2002
  • Nowadays, new product development (NPD) is one of the most crucial factors for business success. The manufacturing firms cannot afford the resources in the long development cycle and the costly redesigns. Good product planning is crucial to ensure the success of NPD, while the Quality Function deployment (QFD) is an effective tool to help the decision makers to determine appropriate product specifications in the product planning stage. Traditionally, in the QFD, the product specifications are determined by a rather subjective evaluation, which is based on the knowledge and experience of the decision makers. In this paper, the traditional QFD methodology is firstly reviewed. An improved Hybrid Quality Function Deployment (HQFD) [MSOfficel] then presented to tackle the shortcomings of traditional QFD methodologies in determining the engineering characteristics. A structured questionnaire to collect and analyze the customer requirements, a methodology to establish a QFD record base and effective case retrieval, and a model to more objectively determine the target values of engineering characteristics are also described.

Stochastics and Artificial Intelligence-based Analytics of Wastewater Plant Operation

  • Sung-Hyun Kwon;Daechul Cho
    • Clean Technology
    • /
    • v.29 no.2
    • /
    • pp.145-150
    • /
    • 2023
  • Tele-metering systems have been useful tools for managing domestic wastewater treatment plants (WWTP) over the last decade. They mostly generate water quality data for discharged water to ensure that it complies with mandatory regulations and they may be able to produce every operation parameter and additional measurements in the near future. A sub-big data group, comprised of about 150,000 data points from four domestic WWTPs, was ready to be classified and also analyzed to optimize the WWTP process. We used the Statistical Product and Service Solutions (SPSS) 25 package in order to statistically treat the data with linear regression and correlation analysis. The major independent variables for analysis were water temperature, sludge recycle rate, electricity used, and water quality of the influent while the dependent variables representing the water quality of the effluent included the total nitrogen, which is the most emphasized index for discharged flow in plants. The water temperature and consumed electricity showed a strong correlation with the total nitrogen but the other indices' mutual correlations with other variables were found to be fuzzy due to the large errors involved. In addition, a multilayer perceptron analysis method was applied to TMS data along with root mean square error (RMSE) analysis. This study showed that the RMSE in the SS, T-N, and TOC predictions were in the range of 10% to 20%.

An Analysis of Design Elements and Satisfaction on the Usability of City Squares - Focused on Gwanghwamun Square and Geumbit Square - (도시광장 설계요소 및 공간이용 만족도 분석 - 광화문광장과 금빛공원광장을 중심으로 -)

  • Choi, Yun Eui;Chon, Jinhyung;Lee, Jung A
    • Journal of the Korean Institute of Landscape Architecture
    • /
    • v.42 no.6
    • /
    • pp.111-123
    • /
    • 2014
  • The city square is an important public open space for people. Being used for various activities, such as community gatherings, open markets, concerts, political rallies, and other events, many types of city squares are represented in a city. Nevertheless, most city squares are planned uniformly, lacking consideration for visitor behavior and use satisfaction. The study investigated the design factors and subcomponents influencing user satisfaction with different types of city squares. This study focused on the general city square in Seoul, including the grand public place (i.e. Gwanghwamun Square) and the neighborhood park (i.e. Geumbit Square). The data were analyzed using factor analysis, linear regression and fuzzy theory. The results of the study are as follows: first, five design factors of satisfaction with city squares are identified (Amenity, Usability, Spatial components, Culture, and Comfortableness). Second, Amenity, Comfortableness, and spatial components significantly affect user satisfaction with Gwanghwamun in that order. On the other hand, in Geumbit Square, Comfortableness, Amenity, Usability and Spatial components affect user satisfaction in a significant way, in that order. Third, cleanliness, a subcomponent of amenity, was ranked highest using the fuzzy theory function for satisfaction with Gwanghwamun Square. Otherwise, the prevalence of plants was ranked the highest on the Geumbit Square survey. The study compared design factors influencing satisfaction in the public grand place and the neighborhood park. The results have implications for designing and planning city squares to the satisfaction of their visitors.

Short-term Forecasting of Power Demand based on AREA (AREA 활용 전력수요 단기 예측)

  • Kwon, S.H.;Oh, H.S.
    • Journal of Korean Society of Industrial and Systems Engineering
    • /
    • v.39 no.1
    • /
    • pp.25-30
    • /
    • 2016
  • It is critical to forecast the maximum daily and monthly demand for power with as little error as possible for our industry and national economy. In general, long-term forecasting of power demand has been studied from both the consumer's perspective and an econometrics model in the form of a generalized linear model with predictors. Time series techniques are used for short-term forecasting with no predictors as predictors must be predicted prior to forecasting response variables and containing estimation errors during this process is inevitable. In previous researches, seasonal exponential smoothing method, SARMA (Seasonal Auto Regressive Moving Average) with consideration to weekly pattern Neuron-Fuzzy model, SVR (Support Vector Regression) model with predictors explored through machine learning, and K-means clustering technique in the various approaches have been applied to short-term power supply forecasting. In this paper, SARMA and intervention model are fitted to forecast the maximum power load daily, weekly, and monthly by using the empirical data from 2011 through 2013. $ARMA(2,\;1,\;2)(1,\;1,\;1)_7$ and $ARMA(0,\;1,\;1)(1,\;1,\;0)_{12}$ are fitted respectively to the daily and monthly power demand, but the weekly power demand is not fitted by AREA because of unit root series. In our fitted intervention model, the factors of long holidays, summer and winter are significant in the form of indicator function. The SARMA with MAPE (Mean Absolute Percentage Error) of 2.45% and intervention model with MAPE of 2.44% are more efficient than the present seasonal exponential smoothing with MAPE of about 4%. Although the dynamic repression model with the predictors of humidity, temperature, and seasonal dummies was applied to foretaste the daily power demand, it lead to a high MAPE of 3.5% even though it has estimation error of predictors.

Effective Drought Prediction Based on Machine Learning (머신러닝 기반 효과적인 가뭄예측)

  • Kim, Kyosik;Yoo, Jae Hwan;Kim, Byunghyun;Han, Kun-Yeun
    • Proceedings of the Korea Water Resources Association Conference
    • /
    • 2021.06a
    • /
    • pp.326-326
    • /
    • 2021
  • 장기간에 걸쳐 넓은 지역에 대해 발생하는 가뭄을 예측하기위해 많은 학자들의 기술적, 학술적 시도가 있어왔다. 본 연구에서는 복잡한 시계열을 가진 가뭄을 전망하는 방법 중 시나리오에 기반을 둔 가뭄전망 방법과 실시간으로 가뭄을 예측하는 비시나리오 기반의 방법 등을 이용하여 미래 가뭄전망을 실시했다. 시나리오에 기반을 둔 가뭄전망 방법으로는, 3개월 GCM(General Circulation Model) 예측 결과를 바탕으로 2009년도 PDSI(Palmer Drought Severity Index) 가뭄지수를 산정하여 가뭄심도에 대한 단기예측을 실시하였다. 또, 통계학적 방법과 물리적 모델(Physical model)에 기반을 둔 확정론적 수치해석 방법을 이용하여 비시나리오 기반 가뭄을 예측했다. 기존 가뭄을 통계학적 방법으로 예측하기 위해서 시도된 대표적인 방법으로 ARIMA(Autoregressive Integrated Moving Average) 모델의 예측에 대한 한계를 극복하기위해 서포트 벡터 회귀(support vector regression, SVR)와 웨이블릿(wavelet neural network) 신경망을 이용해 SPI를 측정하였다. 최적모델구조는 RMSE(root mean square error), MAE(mean absolute error) 및 R(correlation Coefficient)를 통해 선정하였고, 1-6개월의 선행예보 시간을 갖고 가뭄을 전망하였다. 그리고 SPI를 이용하여, 마코프 연쇄(Markov chain) 및 대수선형모델(log-linear model)을 적용하여 SPI기반 가뭄예측의 정확도를 검증하였으며, 터키의 아나톨리아(Anatolia) 지역을 대상으로 뉴로퍼지모델(Neuro-Fuzzy)을 적용하여 1964-2006년 기간의 월평균 강수량과 SPI를 바탕으로 가뭄을 예측하였다. 가뭄 빈도와 패턴이 불규칙적으로 변하며 지역별 강수량의 양극화가 심화됨에 따라 가뭄예측의 정확도를 높여야 하는 요구가 커지고 있다. 본 연구에서는 복잡하고 비선형성으로 이루어진 가뭄 패턴을 기상학적 가뭄의 정도를 나타내는 표준강수증발지수(SPEI, Standardized Precipitation Evapotranspiration Index)인 월SPEI와 일SPEI를 기계학습모델에 적용하여 예측개선 모형을 개발하고자 한다.

  • PDF

Applications of Fuzzy Theory on The Location Decision of Logistics Facilities (퍼지이론을 이용한 물류단지 입지 및 규모결정에 관한 연구)

  • 이승재;정창무;이헌주
    • Journal of Korean Society of Transportation
    • /
    • v.18 no.1
    • /
    • pp.75-85
    • /
    • 2000
  • In existing models in optimization, the crisp data improve has been used in the objective or constraints to derive the optimal solution, Besides, the subjective environments are eliminated because the complex and uncertain circumstances were regarded as Probable ambiguity, In other words those optimal solutions in the existing models could be the complete satisfactory solutions to the objective functions in the Process of application for industrial engineering methods to minimize risks of decision-making. As a result of those, decision-makers in location Problems couldn't face appropriately with the variation of demand as well as other variables and couldn't Provide the chance of wide selection because of the insufficient information. So under the circumstance. it has been to develop the model for the location and size decision problems of logistics facility in the use of the fuzzy theory in the intention of making the most reasonable decision in the Point of subjective view under ambiguous circumstances, in the foundation of the existing decision-making problems which must satisfy the constraints to optimize the objective function in strictly given conditions in this study. Introducing the Process used in this study after the establishment of a general mixed integer Programming(MIP) model based upon the result of existing studies to decide the location and size simultaneously, a fuzzy mixed integer Programming(FMIP) model has been developed in the use of fuzzy theory. And the general linear Programming software, LINDO 6.01 has been used to simulate, to evaluate the developed model with the examples and to judge of the appropriateness and adaptability of the model(FMIP) in the real world.

  • PDF

Dynamic forecasts of bankruptcy with Recurrent Neural Network model (RNN(Recurrent Neural Network)을 이용한 기업부도예측모형에서 회계정보의 동적 변화 연구)

  • Kwon, Hyukkun;Lee, Dongkyu;Shin, Minsoo
    • Journal of Intelligence and Information Systems
    • /
    • v.23 no.3
    • /
    • pp.139-153
    • /
    • 2017
  • Corporate bankruptcy can cause great losses not only to stakeholders but also to many related sectors in society. Through the economic crises, bankruptcy have increased and bankruptcy prediction models have become more and more important. Therefore, corporate bankruptcy has been regarded as one of the major topics of research in business management. Also, many studies in the industry are in progress and important. Previous studies attempted to utilize various methodologies to improve the bankruptcy prediction accuracy and to resolve the overfitting problem, such as Multivariate Discriminant Analysis (MDA), Generalized Linear Model (GLM). These methods are based on statistics. Recently, researchers have used machine learning methodologies such as Support Vector Machine (SVM), Artificial Neural Network (ANN). Furthermore, fuzzy theory and genetic algorithms were used. Because of this change, many of bankruptcy models are developed. Also, performance has been improved. In general, the company's financial and accounting information will change over time. Likewise, the market situation also changes, so there are many difficulties in predicting bankruptcy only with information at a certain point in time. However, even though traditional research has problems that don't take into account the time effect, dynamic model has not been studied much. When we ignore the time effect, we get the biased results. So the static model may not be suitable for predicting bankruptcy. Thus, using the dynamic model, there is a possibility that bankruptcy prediction model is improved. In this paper, we propose RNN (Recurrent Neural Network) which is one of the deep learning methodologies. The RNN learns time series data and the performance is known to be good. Prior to experiment, we selected non-financial firms listed on the KOSPI, KOSDAQ and KONEX markets from 2010 to 2016 for the estimation of the bankruptcy prediction model and the comparison of forecasting performance. In order to prevent a mistake of predicting bankruptcy by using the financial information already reflected in the deterioration of the financial condition of the company, the financial information was collected with a lag of two years, and the default period was defined from January to December of the year. Then we defined the bankruptcy. The bankruptcy we defined is the abolition of the listing due to sluggish earnings. We confirmed abolition of the list at KIND that is corporate stock information website. Then we selected variables at previous papers. The first set of variables are Z-score variables. These variables have become traditional variables in predicting bankruptcy. The second set of variables are dynamic variable set. Finally we selected 240 normal companies and 226 bankrupt companies at the first variable set. Likewise, we selected 229 normal companies and 226 bankrupt companies at the second variable set. We created a model that reflects dynamic changes in time-series financial data and by comparing the suggested model with the analysis of existing bankruptcy predictive models, we found that the suggested model could help to improve the accuracy of bankruptcy predictions. We used financial data in KIS Value (Financial database) and selected Multivariate Discriminant Analysis (MDA), Generalized Linear Model called logistic regression (GLM), Support Vector Machine (SVM), Artificial Neural Network (ANN) model as benchmark. The result of the experiment proved that RNN's performance was better than comparative model. The accuracy of RNN was high in both sets of variables and the Area Under the Curve (AUC) value was also high. Also when we saw the hit-ratio table, the ratio of RNNs that predicted a poor company to be bankrupt was higher than that of other comparative models. However the limitation of this paper is that an overfitting problem occurs during RNN learning. But we expect to be able to solve the overfitting problem by selecting more learning data and appropriate variables. From these result, it is expected that this research will contribute to the development of a bankruptcy prediction by proposing a new dynamic model.