• Title/Summary/Keyword: Statistical optimization

Search Result 649, Processing Time 0.031 seconds

Intelligent Optimal Route Planning Based on Context Awareness (상황인식 기반 지능형 최적 경로계획)

  • Lee, Hyun-Jung;Chang, Yong-Sik
    • Asia pacific journal of information systems
    • /
    • v.19 no.2
    • /
    • pp.117-137
    • /
    • 2009
  • Recently, intelligent traffic information systems have enabled people to forecast traffic conditions before hitting the road. These convenient systems operate on the basis of data reflecting current road and traffic conditions as well as distance-based data between locations. Thanks to the rapid development of ubiquitous computing, tremendous context data have become readily available making vehicle route planning easier than ever. Previous research in relation to optimization of vehicle route planning merely focused on finding the optimal distance between locations. Contexts reflecting the road and traffic conditions were then not seriously treated as a way to resolve the optimal routing problems based on distance-based route planning, because this kind of information does not have much significant impact on traffic routing until a a complex traffic situation arises. Further, it was also not easy to take into full account the traffic contexts for resolving optimal routing problems because predicting the dynamic traffic situations was regarded a daunting task. However, with rapid increase in traffic complexity the importance of developing contexts reflecting data related to moving costs has emerged. Hence, this research proposes a framework designed to resolve an optimal route planning problem by taking full account of additional moving cost such as road traffic cost and weather cost, among others. Recent technological development particularly in the ubiquitous computing environment has facilitated the collection of such data. This framework is based on the contexts of time, traffic, and environment, which addresses the following issues. First, we clarify and classify the diverse contexts that affect a vehicle's velocity and estimates the optimization of moving cost based on dynamic programming that accounts for the context cost according to the variance of contexts. Second, the velocity reduction rate is applied to find the optimal route (shortest path) using the context data on the current traffic condition. The velocity reduction rate infers to the degree of possible velocity including moving vehicles' considerable road and traffic contexts, indicating the statistical or experimental data. Knowledge generated in this papercan be referenced by several organizations which deal with road and traffic data. Third, in experimentation, we evaluate the effectiveness of the proposed context-based optimal route (shortest path) between locations by comparing it to the previously used distance-based shortest path. A vehicles' optimal route might change due to its diverse velocity caused by unexpected but potential dynamic situations depending on the road condition. This study includes such context variables as 'road congestion', 'work', 'accident', and 'weather' which can alter the traffic condition. The contexts can affect moving vehicle's velocity on the road. Since these context variables except for 'weather' are related to road conditions, relevant data were provided by the Korea Expressway Corporation. The 'weather'-related data were attained from the Korea Meteorological Administration. The aware contexts are classified contexts causing reduction of vehicles' velocity which determines the velocity reduction rate. To find the optimal route (shortest path), we introduced the velocity reduction rate in the context for calculating a vehicle's velocity reflecting composite contexts when one event synchronizes with another. We then proposed a context-based optimal route (shortest path) algorithm based on the dynamic programming. The algorithm is composed of three steps. In the first initialization step, departure and destination locations are given, and the path step is initialized as 0. In the second step, moving costs including composite contexts into account between locations on path are estimated using the velocity reduction rate by context as increasing path steps. In the third step, the optimal route (shortest path) is retrieved through back-tracking. In the provided research model, we designed a framework to account for context awareness, moving cost estimation (taking both composite and single contexts into account), and optimal route (shortest path) algorithm (based on dynamic programming). Through illustrative experimentation using the Wilcoxon signed rank test, we proved that context-based route planning is much more effective than distance-based route planning., In addition, we found that the optimal solution (shortest paths) through the distance-based route planning might not be optimized in real situation because road condition is very dynamic and unpredictable while affecting most vehicles' moving costs. For further study, while more information is needed for a more accurate estimation of moving vehicles' costs, this study still stands viable in the applications to reduce moving costs by effective route planning. For instance, it could be applied to deliverers' decision making to enhance their decision satisfaction when they meet unpredictable dynamic situations in moving vehicles on the road. Overall, we conclude that taking into account the contexts as a part of costs is a meaningful and sensible approach to in resolving the optimal route problem.

Teachers' Recognition on the Optimization of the Educational Contents of Clothing and Textiles in Practical Arts or Technology.Home Economics (실과 및 기술.가정 교과에서 의생활 교육내용의 적정성에 대한 교사의 인식)

  • Baek Seung-Hee;Han Young-Sook;Lee Hye-Ja
    • Journal of Korean Home Economics Education Association
    • /
    • v.18 no.3 s.41
    • /
    • pp.97-117
    • /
    • 2006
  • The purpose of this study was to investigate the teachers' recognition on the optimization of the educational contents of Clothing & Textiles in subjects of :he Practical Arts or the Technology & Home Economics in the course of elementary, middle and high schools. The statistical data for this research were collected from 203 questionnaires of teachers who work on elementary, middle and high schools. Mean. standard deviation, percentage were calculated using SPSS/WIN 12.0 program. Also. these materials were verified by t-test, One-way ANOVA and post verification Duncan. The results were as follows; First, The equipment ratio of practice laboratory were about 24% and very poor in elementary schools but those of middle and high school were 97% and 78% each and higher than elementary schools. Second, More than 50% of teachers recognized the amount of learning 'proper'. The elementary school teachers recognized the mount of learning in 'operating sewing machines' too heavy especially, the same as middle school teachers in 'making shorts': the same as high school teachers in 'making tablecloth and curtain' and 'making pillow cover or bag'. Third, All of the elementary, middle and high school teachers recognized the levels of total contents of clothing and textiles 'common'. The 80% of elementary school teachers recognized 'operating sewing machines' and 'making cushions' difficult especially. The same as middle school teachers in 'hand knitting handbag by crochet hoop needle', 'the various kinds of cloth' and 'making short pants'. The same as high school teachers in 'making tablecloth or curtain'. Fourth, Elementary school teachers recognized 'practicing basic hand needlework' and 'making pouch using hand needlework' important in the degree of educational contents importance. Middle school teachers recognized 'making short pants unimportant. High school teachers considered the contents focusing on practice such as 'making tablecloth and curtain' and 'making pillow cover or bags' unimportant. My suggestions were as follows; Both laboratories and facilities for practice should be established for making clothing and textiles lessons effective in Practical Arts in elementary schools. The 'operating sewing machines' which were considered difficult should be dealt in upper grade, re-conditioning to easier or omitted. The practical contents should be changed to student-activity-oriented and should be recomposed in order to familiar with students' living. It was needed to various and sufficient supports for increasing the teachers' practical abilities.

  • PDF

Optimization of Extraction Conditions to Obtain Functional Components from Buckwheat (Fagopyrum esculentum M.) Sprouts, using Response Surface Methodology (반응표면분석법에 의한 메밀(Fagopyrum esculentum M.) 새싹 기능성분의 추출 조건 최적화)

  • Park, Kee-Jai;Lim, Jeong-Ho;Kim, Bum-Keun;Jeong, Jin-Woong;Kim, Jong-Chan;Lee, Myung-Heon;Cho, Young-Sim;Jung, Hee-Yong
    • Food Science and Preservation
    • /
    • v.16 no.5
    • /
    • pp.734-741
    • /
    • 2009
  • Response surface methodology (RSM) was used to optimize extraction conditions for functional components of buckwheat (Fagopyrum esculentum). A central composite design was applied to investigate the effects of three independent variables, namelyextraction temperature (X1), extraction time (X2), and ethanol concentration (X3), on responses including extraction yield (Y1), total phenolic content in the extract (Y2), $\alpha$-glucosidase inhibition activity (Y3), and acetylcholine esterase (ACE) inhibition activity (Y4). Data were analyzed using an expert design strategy and statistical software. The maximum yield was 24.95% (w/w) at $55.75^{\circ}C$ extraction temperature, 8.75 hextraction time, and 15.65% (v/v) ethanol. The maximum total phenolic yield was 222.45 mg/100 g under the conditions of $28.11^{\circ}C$ extraction temperature, 8.65 h extraction time, and 81.72% (v/v) ethanol. The maximum $\alpha$-glucosidase inhibition activity was 85.38% at $9.62^{\circ}C$, 7.86 h, and 57.58% (v/v) ethanol. The maximum ACE inhibition activity was 86.91% under extraction conditions of $10.12^{\circ}C$, 4.86 h, and 44.44% (v/v) ethanol. Based on superimposition of a four-dimensional RSM with respect to levels of total phenolics, $\alpha$-glucosidase inhibition activity, and ACE inhibition activity, obtained under various extraction conditions, the optimum ranges of conditions were an extraction temperature of $0-70^{\circ}C$, an extraction time of 2-8 h, and an ethanol concentration of 30-80% (v/v).

Optimization of TDA Recycling Process for TDI Residue using Near-critical Hydrolysis Process (근임계수 가수분해 공정을 이용한 TDI 공정 폐기물로부터 TDA 회수 공정 최적화)

  • Han, Joo Hee;Han, Kee Do;Jeong, Chang Mo;Do, Seung Hoe;Sin, Yeong Ho
    • Korean Chemical Engineering Research
    • /
    • v.44 no.6
    • /
    • pp.650-658
    • /
    • 2006
  • The recycling of TDA from solid waste of TDI plant(TDI-R) by near-critical hydrolysis reaction had been studied by means of a statistical design of experiment. The main and interaction effects of process variables had been defined from the experiments in a batch reactor and the correlation equation with process variables for TDA yield had been obtained from the experiments in a continuous pilot plant. It was confirmed that the effects of reaction temperature, catalyst type and concentration, and the weight ratio of water to TDI-R(WR) on TDA yield were significant. TDA yield decreased with increases in reaction temperature and catalyst concentration, and increased with an increase in WR. As a catalyst, NaOH was more effective than $Na_2CO_3$ for TDA yield. The interaction effects between catalyst concentration and temperature, WR and temperature, catalyst type and reaction time on TDA yield had been defined as significant. Although the effect of catalyst concentration on TDA yield at $300^{\circ}C$ as subcritical water was insignificant, the TDA yield decreased with increasing catalyst concentration at $400^{\circ}C$ as supercritical water. On the other hand, the yield increased with an increase in WR at $300^{\circ}C$ but showed negligible effect with WR at $400^{\circ}C$. The optimization of process variables for TDA yield has been explored with a pilot plant for scale-up. The catalyst concentration and WR were selected as process variables with respect to economic feasibility and efficiency. The effects of process variables on TDA yield had been explored by means of central composite design. The TDA yield increased with an increase in catalyst concentration. It showed maximum value at below 2.5 of WR and then decreased with an increase in WR. However, the ratio at which the TDA yield showed a maximum value increased with increasing catalyst concentration. The correlation equation of a quadratic model with catalyst concentration and WR had been obtained by the regression analysis of experimental results in a pilot plant.

Optimization of Medium for the Carotenoid Production by Rhodobacter sphaeroides PS-24 Using Response Surface Methodology (반응 표면 분석법을 사용한 Rhodobacter sphaeroides PS-24 유래 carotenoid 생산 배지 최적화)

  • Bong, Ki-Moon;Kim, Kong-Min;Seo, Min-Kyoung;Han, Ji-Hee;Park, In-Chul;Lee, Chul-Won;Kim, Pyoung-Il
    • Korean Journal of Organic Agriculture
    • /
    • v.25 no.1
    • /
    • pp.135-148
    • /
    • 2017
  • Response Surface Methodology (RSM), which is combining with Plackett-Burman design and Box-Behnken experimental design, was applied to optimize the ratios of the nutrient components for carotenoid production by Rhodobacter sphaeroides PS-24 in liquid state fermentation. Nine nutrient ingredients containing yeast extract, sodium acetate, NaCl, $K_2HPO_4$, $MgSO_4$, mono-sodium glutamate, $Na_2CO_3$, $NH_4Cl$ and $CaCl_2$ were finally selected for optimizing the medium composition based on their statistical significance and positive effects on carotenoid yield. Box-Behnken design was employed for further optimization of the selected nutrient components in order to increase carotenoid production. Based on the Box-Behnken assay data, the secondary order coefficient model was set up to investigate the relationship between the carotenoid productivity and nutrient ingredients. The important factors having influence on optimal medium constituents for carotenoid production by Rhodobacter sphaeroides PS-24 were determined as follows: yeast extract 1.23 g, sodium acetate 1 g, $NH_4Cl$ 1.75 g, NaCl 2.5 g, $K_2HPO_4$ 2 g, $MgSO_4$ 1.0 g, mono-sodium glutamate 7.5 g, $Na_2CO_3$ 3.71 g, $NH_4Cl$ 3.5g, $CaCl_2$ 0.01 g, per liter. Maximum carotenoid yield of 18.11 mg/L was measured by confirmatory experiment in liquid culture using 500 L fermenter.

Optimization of the Indole-3-Acetic Acid Production Medium of Pantoea agglomerans SRCM 119864 using Response Surface Methodology (반응표면분석법을 활용한 Pantoea agglomerans SRCM 119864의 Indole-3-acetic acid 생산 배지 최적화)

  • Ho Jin, Jeong;Gwangsu, Ha;Su Ji, Jeong;Myeong Seon, Ryu;JinWon, Kim;Do-Youn, Jeong;Hee-Jong, Yang
    • Journal of Life Science
    • /
    • v.32 no.11
    • /
    • pp.872-881
    • /
    • 2022
  • In this study, we optimized the composition of the indole-3-acetic acid (IAA) production medium using response surface methodology on Pantoea agglomerans SRCM 119864 isolated from soil. IAA-producing P. aglomerans SRCM 119864 was identified by 16S rRNA gene sequencing. There are 11 intermediate components known to affect IAA production, hence the effect of each component on IAA production was investigated using a Plackett-Burman design (PBD). Based on the PBD, sucrose, tryptone, and sodium chloride were selected as the main factors that enhanced the IAA production at optimal L-tryptophan concentration. The predicted maximum IAA production (64.34 mg/l) was obtained for a concentration of sucrose of 13.38 g/l, of tryptone of 18.34 g/l, of sodium chloride of 9.71 g/l, and of L-tryptophan of 6.25 g/l using a the hybrid design experimental model. In the experiment, the nutrient broth medium supplemented with 0.1% L-tryptophan as the basal medium produced 45.24 mg/l of IAA, whereas the optimized medium produced 65.40 mg/l of IAA, resulting in a 44.56% increase in efficiency. It was confirmed that the IAA production of the designed optimal composition medium was very similar to the predicted IAA production. The statistical significance and suitability of the experimental model were verified through analysis of variance (ANOVA). Therefore, in this study, we determined the optimal growth medium concentration for the maximum production of IAA, which can contribute to sustainable agriculture and increase crop yield.

Social Network-based Hybrid Collaborative Filtering using Genetic Algorithms (유전자 알고리즘을 활용한 소셜네트워크 기반 하이브리드 협업필터링)

  • Noh, Heeryong;Choi, Seulbi;Ahn, Hyunchul
    • Journal of Intelligence and Information Systems
    • /
    • v.23 no.2
    • /
    • pp.19-38
    • /
    • 2017
  • Collaborative filtering (CF) algorithm has been popularly used for implementing recommender systems. Until now, there have been many prior studies to improve the accuracy of CF. Among them, some recent studies adopt 'hybrid recommendation approach', which enhances the performance of conventional CF by using additional information. In this research, we propose a new hybrid recommender system which fuses CF and the results from the social network analysis on trust and distrust relationship networks among users to enhance prediction accuracy. The proposed algorithm of our study is based on memory-based CF. But, when calculating the similarity between users in CF, our proposed algorithm considers not only the correlation of the users' numeric rating patterns, but also the users' in-degree centrality values derived from trust and distrust relationship networks. In specific, it is designed to amplify the similarity between a target user and his or her neighbor when the neighbor has higher in-degree centrality in the trust relationship network. Also, it attenuates the similarity between a target user and his or her neighbor when the neighbor has higher in-degree centrality in the distrust relationship network. Our proposed algorithm considers four (4) types of user relationships - direct trust, indirect trust, direct distrust, and indirect distrust - in total. And, it uses four adjusting coefficients, which adjusts the level of amplification / attenuation for in-degree centrality values derived from direct / indirect trust and distrust relationship networks. To determine optimal adjusting coefficients, genetic algorithms (GA) has been adopted. Under this background, we named our proposed algorithm as SNACF-GA (Social Network Analysis - based CF using GA). To validate the performance of the SNACF-GA, we used a real-world data set which is called 'Extended Epinions dataset' provided by 'trustlet.org'. It is the data set contains user responses (rating scores and reviews) after purchasing specific items (e.g. car, movie, music, book) as well as trust / distrust relationship information indicating whom to trust or distrust between users. The experimental system was basically developed using Microsoft Visual Basic for Applications (VBA), but we also used UCINET 6 for calculating the in-degree centrality of trust / distrust relationship networks. In addition, we used Palisade Software's Evolver, which is a commercial software implements genetic algorithm. To examine the effectiveness of our proposed system more precisely, we adopted two comparison models. The first comparison model is conventional CF. It only uses users' explicit numeric ratings when calculating the similarities between users. That is, it does not consider trust / distrust relationship between users at all. The second comparison model is SNACF (Social Network Analysis - based CF). SNACF differs from the proposed algorithm SNACF-GA in that it considers only direct trust / distrust relationships. It also does not use GA optimization. The performances of the proposed algorithm and comparison models were evaluated by using average MAE (mean absolute error). Experimental result showed that the optimal adjusting coefficients for direct trust, indirect trust, direct distrust, indirect distrust were 0, 1.4287, 1.5, 0.4615 each. This implies that distrust relationships between users are more important than trust ones in recommender systems. From the perspective of recommendation accuracy, SNACF-GA (Avg. MAE = 0.111943), the proposed algorithm which reflects both direct and indirect trust / distrust relationships information, was found to greatly outperform a conventional CF (Avg. MAE = 0.112638). Also, the algorithm showed better recommendation accuracy than the SNACF (Avg. MAE = 0.112209). To confirm whether these differences are statistically significant or not, we applied paired samples t-test. The results from the paired samples t-test presented that the difference between SNACF-GA and conventional CF was statistical significant at the 1% significance level, and the difference between SNACF-GA and SNACF was statistical significant at the 5%. Our study found that the trust/distrust relationship can be important information for improving performance of recommendation algorithms. Especially, distrust relationship information was found to have a greater impact on the performance improvement of CF. This implies that we need to have more attention on distrust (negative) relationships rather than trust (positive) ones when tracking and managing social relationships between users.

The Analysis on the Relationship between Firms' Exposures to SNS and Stock Prices in Korea (기업의 SNS 노출과 주식 수익률간의 관계 분석)

  • Kim, Taehwan;Jung, Woo-Jin;Lee, Sang-Yong Tom
    • Asia pacific journal of information systems
    • /
    • v.24 no.2
    • /
    • pp.233-253
    • /
    • 2014
  • Can the stock market really be predicted? Stock market prediction has attracted much attention from many fields including business, economics, statistics, and mathematics. Early research on stock market prediction was based on random walk theory (RWT) and the efficient market hypothesis (EMH). According to the EMH, stock market are largely driven by new information rather than present and past prices. Since it is unpredictable, stock market will follow a random walk. Even though these theories, Schumaker [2010] asserted that people keep trying to predict the stock market by using artificial intelligence, statistical estimates, and mathematical models. Mathematical approaches include Percolation Methods, Log-Periodic Oscillations and Wavelet Transforms to model future prices. Examples of artificial intelligence approaches that deals with optimization and machine learning are Genetic Algorithms, Support Vector Machines (SVM) and Neural Networks. Statistical approaches typically predicts the future by using past stock market data. Recently, financial engineers have started to predict the stock prices movement pattern by using the SNS data. SNS is the place where peoples opinions and ideas are freely flow and affect others' beliefs on certain things. Through word-of-mouth in SNS, people share product usage experiences, subjective feelings, and commonly accompanying sentiment or mood with others. An increasing number of empirical analyses of sentiment and mood are based on textual collections of public user generated data on the web. The Opinion mining is one domain of the data mining fields extracting public opinions exposed in SNS by utilizing data mining. There have been many studies on the issues of opinion mining from Web sources such as product reviews, forum posts and blogs. In relation to this literatures, we are trying to understand the effects of SNS exposures of firms on stock prices in Korea. Similarly to Bollen et al. [2011], we empirically analyze the impact of SNS exposures on stock return rates. We use Social Metrics by Daum Soft, an SNS big data analysis company in Korea. Social Metrics provides trends and public opinions in Twitter and blogs by using natural language process and analysis tools. It collects the sentences circulated in the Twitter in real time, and breaks down these sentences into the word units and then extracts keywords. In this study, we classify firms' exposures in SNS into two groups: positive and negative. To test the correlation and causation relationship between SNS exposures and stock price returns, we first collect 252 firms' stock prices and KRX100 index in the Korea Stock Exchange (KRX) from May 25, 2012 to September 1, 2012. We also gather the public attitudes (positive, negative) about these firms from Social Metrics over the same period of time. We conduct regression analysis between stock prices and the number of SNS exposures. Having checked the correlation between the two variables, we perform Granger causality test to see the causation direction between the two variables. The research result is that the number of total SNS exposures is positively related with stock market returns. The number of positive mentions of has also positive relationship with stock market returns. Contrarily, the number of negative mentions has negative relationship with stock market returns, but this relationship is statistically not significant. This means that the impact of positive mentions is statistically bigger than the impact of negative mentions. We also investigate whether the impacts are moderated by industry type and firm's size. We find that the SNS exposures impacts are bigger for IT firms than for non-IT firms, and bigger for small sized firms than for large sized firms. The results of Granger causality test shows change of stock price return is caused by SNS exposures, while the causation of the other way round is not significant. Therefore the correlation relationship between SNS exposures and stock prices has uni-direction causality. The more a firm is exposed in SNS, the more is the stock price likely to increase, while stock price changes may not cause more SNS mentions.

A Study on the Prediction Model of Stock Price Index Trend based on GA-MSVM that Simultaneously Optimizes Feature and Instance Selection (입력변수 및 학습사례 선정을 동시에 최적화하는 GA-MSVM 기반 주가지수 추세 예측 모형에 관한 연구)

  • Lee, Jong-sik;Ahn, Hyunchul
    • Journal of Intelligence and Information Systems
    • /
    • v.23 no.4
    • /
    • pp.147-168
    • /
    • 2017
  • There have been many studies on accurate stock market forecasting in academia for a long time, and now there are also various forecasting models using various techniques. Recently, many attempts have been made to predict the stock index using various machine learning methods including Deep Learning. Although the fundamental analysis and the technical analysis method are used for the analysis of the traditional stock investment transaction, the technical analysis method is more useful for the application of the short-term transaction prediction or statistical and mathematical techniques. Most of the studies that have been conducted using these technical indicators have studied the model of predicting stock prices by binary classification - rising or falling - of stock market fluctuations in the future market (usually next trading day). However, it is also true that this binary classification has many unfavorable aspects in predicting trends, identifying trading signals, or signaling portfolio rebalancing. In this study, we try to predict the stock index by expanding the stock index trend (upward trend, boxed, downward trend) to the multiple classification system in the existing binary index method. In order to solve this multi-classification problem, a technique such as Multinomial Logistic Regression Analysis (MLOGIT), Multiple Discriminant Analysis (MDA) or Artificial Neural Networks (ANN) we propose an optimization model using Genetic Algorithm as a wrapper for improving the performance of this model using Multi-classification Support Vector Machines (MSVM), which has proved to be superior in prediction performance. In particular, the proposed model named GA-MSVM is designed to maximize model performance by optimizing not only the kernel function parameters of MSVM, but also the optimal selection of input variables (feature selection) as well as instance selection. In order to verify the performance of the proposed model, we applied the proposed method to the real data. The results show that the proposed method is more effective than the conventional multivariate SVM, which has been known to show the best prediction performance up to now, as well as existing artificial intelligence / data mining techniques such as MDA, MLOGIT, CBR, and it is confirmed that the prediction performance is better than this. Especially, it has been confirmed that the 'instance selection' plays a very important role in predicting the stock index trend, and it is confirmed that the improvement effect of the model is more important than other factors. To verify the usefulness of GA-MSVM, we applied it to Korea's real KOSPI200 stock index trend forecast. Our research is primarily aimed at predicting trend segments to capture signal acquisition or short-term trend transition points. The experimental data set includes technical indicators such as the price and volatility index (2004 ~ 2017) and macroeconomic data (interest rate, exchange rate, S&P 500, etc.) of KOSPI200 stock index in Korea. Using a variety of statistical methods including one-way ANOVA and stepwise MDA, 15 indicators were selected as candidate independent variables. The dependent variable, trend classification, was classified into three states: 1 (upward trend), 0 (boxed), and -1 (downward trend). 70% of the total data for each class was used for training and the remaining 30% was used for verifying. To verify the performance of the proposed model, several comparative model experiments such as MDA, MLOGIT, CBR, ANN and MSVM were conducted. MSVM has adopted the One-Against-One (OAO) approach, which is known as the most accurate approach among the various MSVM approaches. Although there are some limitations, the final experimental results demonstrate that the proposed model, GA-MSVM, performs at a significantly higher level than all comparative models.

Ensemble Learning with Support Vector Machines for Bond Rating (회사채 신용등급 예측을 위한 SVM 앙상블학습)

  • Kim, Myoung-Jong
    • Journal of Intelligence and Information Systems
    • /
    • v.18 no.2
    • /
    • pp.29-45
    • /
    • 2012
  • Bond rating is regarded as an important event for measuring financial risk of companies and for determining the investment returns of investors. As a result, it has been a popular research topic for researchers to predict companies' credit ratings by applying statistical and machine learning techniques. The statistical techniques, including multiple regression, multiple discriminant analysis (MDA), logistic models (LOGIT), and probit analysis, have been traditionally used in bond rating. However, one major drawback is that it should be based on strict assumptions. Such strict assumptions include linearity, normality, independence among predictor variables and pre-existing functional forms relating the criterion variablesand the predictor variables. Those strict assumptions of traditional statistics have limited their application to the real world. Machine learning techniques also used in bond rating prediction models include decision trees (DT), neural networks (NN), and Support Vector Machine (SVM). Especially, SVM is recognized as a new and promising classification and regression analysis method. SVM learns a separating hyperplane that can maximize the margin between two categories. SVM is simple enough to be analyzed mathematical, and leads to high performance in practical applications. SVM implements the structuralrisk minimization principle and searches to minimize an upper bound of the generalization error. In addition, the solution of SVM may be a global optimum and thus, overfitting is unlikely to occur with SVM. In addition, SVM does not require too many data sample for training since it builds prediction models by only using some representative sample near the boundaries called support vectors. A number of experimental researches have indicated that SVM has been successfully applied in a variety of pattern recognition fields. However, there are three major drawbacks that can be potential causes for degrading SVM's performance. First, SVM is originally proposed for solving binary-class classification problems. Methods for combining SVMs for multi-class classification such as One-Against-One, One-Against-All have been proposed, but they do not improve the performance in multi-class classification problem as much as SVM for binary-class classification. Second, approximation algorithms (e.g. decomposition methods, sequential minimal optimization algorithm) could be used for effective multi-class computation to reduce computation time, but it could deteriorate classification performance. Third, the difficulty in multi-class prediction problems is in data imbalance problem that can occur when the number of instances in one class greatly outnumbers the number of instances in the other class. Such data sets often cause a default classifier to be built due to skewed boundary and thus the reduction in the classification accuracy of such a classifier. SVM ensemble learning is one of machine learning methods to cope with the above drawbacks. Ensemble learning is a method for improving the performance of classification and prediction algorithms. AdaBoost is one of the widely used ensemble learning techniques. It constructs a composite classifier by sequentially training classifiers while increasing weight on the misclassified observations through iterations. The observations that are incorrectly predicted by previous classifiers are chosen more often than examples that are correctly predicted. Thus Boosting attempts to produce new classifiers that are better able to predict examples for which the current ensemble's performance is poor. In this way, it can reinforce the training of the misclassified observations of the minority class. This paper proposes a multiclass Geometric Mean-based Boosting (MGM-Boost) to resolve multiclass prediction problem. Since MGM-Boost introduces the notion of geometric mean into AdaBoost, it can perform learning process considering the geometric mean-based accuracy and errors of multiclass. This study applies MGM-Boost to the real-world bond rating case for Korean companies to examine the feasibility of MGM-Boost. 10-fold cross validations for threetimes with different random seeds are performed in order to ensure that the comparison among three different classifiers does not happen by chance. For each of 10-fold cross validation, the entire data set is first partitioned into tenequal-sized sets, and then each set is in turn used as the test set while the classifier trains on the other nine sets. That is, cross-validated folds have been tested independently of each algorithm. Through these steps, we have obtained the results for classifiers on each of the 30 experiments. In the comparison of arithmetic mean-based prediction accuracy between individual classifiers, MGM-Boost (52.95%) shows higher prediction accuracy than both AdaBoost (51.69%) and SVM (49.47%). MGM-Boost (28.12%) also shows the higher prediction accuracy than AdaBoost (24.65%) and SVM (15.42%)in terms of geometric mean-based prediction accuracy. T-test is used to examine whether the performance of each classifiers for 30 folds is significantly different. The results indicate that performance of MGM-Boost is significantly different from AdaBoost and SVM classifiers at 1% level. These results mean that MGM-Boost can provide robust and stable solutions to multi-classproblems such as bond rating.