Browse > Article
http://dx.doi.org/10.5351/CKSS.2008.15.1.043

Sparse Multinomial Kernel Logistic Regression  

Shim, Joo-Yong (Department of Applied Statistics, Catholic University of Daegu)
Bae, Jong-Sig (Department of Mathematics, Sungkyunkwan University)
Hwang, Chang-Ha (Division of Information and Computer Science, Dankook University)
Publication Information
Communications for Statistical Applications and Methods / v.15, no.1, 2008 , pp. 43-50 More about this Journal
Abstract
Multinomial logistic regression is a well known multiclass classification method in the field of statistical learning. More recently, the development of sparse multinomial logistic regression model has found application in microarray classification, where explicit identification of the most informative observations is of value. In this paper, we propose a sparse multinomial kernel logistic regression model, in which the sparsity arises from the use of a Laplacian prior and a fast exact algorithm is derived by employing a bound optimization approach. Experimental results are then presented to indicate the performance of the proposed procedure.
Keywords
Bound optimization; Laplacian regularization; multinomial logistic regression; sparsity; support vector machine;
Citations & Related Records
연도 인용수 순위
  • Reference
1 Krishnapuram, B., Carin, L., Figueiredo, M. A. T. and Hartemink, A. J. (2005). Sparse multi-nomial logistic regression: fast algorithms and generalization bounds. IEEE Ttransaction on Pattern Analysis and Machine Intelligence, 27, 957-968   DOI   ScienceOn
2 Rifkin, R. and Klautau, A. (2004). In defense of one-vs-all classification. Journal of Machine Learning Research, 5, 101-141
3 Minka, T. (2003). A comparison of numerical optimizers for logistic regression. Technical Report, Department of Statistics, Carnegie Mellon University
4 Lawrence, N. D., Seeger, M. and Herbrich, R. (2003). Fast sparse Gaussian process methods: the informative vector machine. Advances in Neural Information Processing Systems, 15, 609-616
5 Mercer, J. (1909). Functions of positive and negative type and their connection with the theory of integral equations. Philosophical Transactions of the Royal Society of London, 209, 415-446   DOI
6 Csato, L. and Opper, M. (2002). Sparse online Gaussian processes. Neural Computation, 14, 641-668   DOI   ScienceOn
7 Kimeldorf, G. S. and Wahba, G. (1971). Some results on Tchebycheffian spline functions. Journal of Mathematical Analysis and its Applications, 33, 82-95   DOI
8 Bohning, D. (1992). Multinomial logistic regression algorithm. Annals of the Institute of Statistical Mathematics, 44, 197-200   DOI
9 Tipping, M. (2001). Sparse Bayesian learning and the relevance vector machine. Journal of Machine Learning Research, 1, 211-244   DOI
10 Vapnik, V. N. (1995). The Nature of Statistical Learning Theory. Springer-Verlag, New York
11 Cawley, G. C., Talbot, N. L. C. and Girolami, M. (2006). Sparse multinomial logistic regression via Bayesian L1 regularisation. Advances in Neural Information Processing Systems, 18, 609-616