Browse > Article
http://dx.doi.org/10.6109/jkiice.2010.14.1.063

An Empirical Analysis of Boosing of Neural Networks for Bankruptcy Prediction  

Kim, Myoung-Jong (동서대학교 경영학부)
Kang, Dae-Ki (동서대학교 컴퓨터정보공학부)
Abstract
Ensemble is one of widely used methods for improving the performance of classification and prediction models. Two popular ensemble methods, Bagging and Boosting, have been applied with great success to various machine learning problems using mostly decision trees as base classifiers. This paper performs an empirical comparison of Boosted neural networks and traditional neural networks on bankruptcy prediction tasks. Experimental results on Korean firms indicated that the boosted neural networks showed the improved performance over traditional neural networks.
Keywords
Boosting; Neural Networks; Bankruptcy Prediction;
Citations & Related Records
연도 인용수 순위
  • Reference
1 Evgeniou, T., Perez-Breva, L., Pontil, M., & Poggio, T. (2000). Bound on the generalization performance of kernel machine ensembles, Proc. ICMI, 271-278.
2 Bauer, E., & Kohavi, R. (1999). An empirical comparison of voting classification algorithms: Bagging, boosting, and variants. Machine Learning, 36, 105-139.   DOI   ScienceOn
3 Breiman, L. (1994). Bagging predictors. Machine Learning, 24(2), 123-140.
4 Dong, Y. S., & Han, K. S. (2004). A comparison of several ensemble methods for text categorization, IEEE International Conference on Service Computing.
5 Fawcett, T. (2006). An introduction to ROC analysis. Pattern Recognition Letters, 27, 861-874.   DOI   ScienceOn
6 Freund, Y., & Schapire, R E. (1996). Experiments with a new boosting algorithm. Machine Learning: Proceedings of Thirteenth International Conference, 148-156.
7 Kim, H. C., Pang, S., Je, H. M., K, D. J., & Bang, S. Y. (2003). Constructing support vector machine ensemble, Pattern Recognition, 36, 2757-2767.   DOI   ScienceOn
8 Alfaro, E., Gamez, M., & Garcia, N. (2007). Multiclass corporate failure prediction by AdaBoost.M1. Advanced Economic Research, 13, 301-312.   DOI   ScienceOn
9 Alfaro, E., Garcia, N., Gamez, M., & Elizondo, D. (2008). Bankruptcy forecasting: an empirical comparison of AdaBooost and neural networks. Decision Support Systems, 45, 110-122.   DOI   ScienceOn
10 Maclin, R, & Opitz, D. (1997). An empirical evaluation of bagging and boosting. In Proceedings of the Fourteenth National Conference on Artificial Intelligence, 546-551.
11 Quinlan, J. R. (1996). Bagging, boosting and C4.5. Machine Learning: Proceedings of the Fourteenth International Conference, 725-730.
12 Freund, Y. (1995). Boosting a weak learning algorithm by majority. Information and Computation, 121(2), 256-285.   DOI   ScienceOn
13 Optiz, D., & Maclin, R (1999). Popular ensemble methods: an empirical study. Journal of Artificial Intelligence, 11, 169-198.
14 Freund, Y., & Schapire, R. E. (1997). A decision theoretic generalization of online learning and an application to boosting. Journal of Computer and System Science, 55(1), 119-139.   DOI   ScienceOn
15 Breiman, L. (1996). Bias, variance, and arcing classifiers (Tech.Rep.No.460). Berkeley: Statistics Department, University of California at Berkeley.
16 Hansen, L., & Salamon, P. (1990). Neural network ensembles, IEEE Trans. PAMI, Vol.12, 993-1001.
17 Drucker, H., & Cortes, C. (1996). Boosting decision trees, Advanced Neural Information Processing Systems, 8.
18 Buciu, I., Kotrooulos, C., & Pitas, I. (2001). Combining support vector machines for accuracy face detection, Proc. ICIP, 1054-1057.