• Title/Summary/Keyword: Support Vector Machines(SVM)

Search Result 282, Processing Time 0.022 seconds

Spam Filter by Using X2 Statistics and Support Vector Machines (카이제곱 통계량과 지지벡터기계를 이용한 스팸메일 필터)

  • Lee, Song-Wook
    • The KIPS Transactions:PartB
    • /
    • v.17B no.3
    • /
    • pp.249-254
    • /
    • 2010
  • We propose an automatic spam filter for e-mail data using Support Vector Machines(SVM). We use a lexical form of a word and its part of speech(POS) tags as features and select features by chi square statistics. We represent each feature by TF(text frequency), TF-IDF, and binary weight for experiments. After training SVM with the selected features, SVM classifies each e-mail as spam or not. In experiment, the selected features improve the performance of our system and we acquired overall 98.9% of accuracy with TREC05-p1 spam corpus.

An Early Warning Model for Student Status Based on Genetic Algorithm-Optimized Radial Basis Kernel Support Vector Machine

  • Hui Li;Qixuan Huang;Chao Wang
    • Journal of Information Processing Systems
    • /
    • v.20 no.2
    • /
    • pp.263-272
    • /
    • 2024
  • A model based on genetic algorithm optimization, GA-SVM, is proposed to warn university students of their status. This model improves the predictive effect of support vector machines. The genetic optimization algorithm is used to train the hyperparameters and adjust the kernel parameters, kernel penalty factor C, and gamma to optimize the support vector machine model, which can rapidly achieve convergence to obtain the optimal solution. The experimental model was trained on open-source datasets and validated through comparisons with random forest, backpropagation neural network, and GA-SVM models. The test results show that the genetic algorithm-optimized radial basis kernel support vector machine model GA-SVM can obtain higher accuracy rates when used for early warning in university learning.

Customer Level Classification Model Using Ordinal Multiclass Support Vector Machines

  • Kim, Kyoung-Jae;Ahn, Hyun-Chul
    • Asia pacific journal of information systems
    • /
    • v.20 no.2
    • /
    • pp.23-37
    • /
    • 2010
  • Conventional Support Vector Machines (SVMs) have been utilized as classifiers for binary classification problems. However, certain real world problems, including corporate bond rating, cannot be addressed by binary classifiers because these are multi-class problems. For this reason, numerous studies have attempted to transform the original SVM into a multiclass classifier. These studies, however, have only considered nominal classification problems. Thus, these approaches have been limited by the existence of multiclass classification problems where classes are not nominal but ordinal in real world, such as corporate bond rating and multiclass customer classification. In this study, we adopt a novel multiclass SVM which can address ordinal classification problems using ordinal pairwise partitioning (OPP). The proposed model in our study may use fewer classifiers, but it classifies more accurately because it considers the characteristics of the order of the classes. Although it can be applied to all kinds of ordinal multiclass classification problems, most prior studies have applied it to finance area like bond rating. Thus, this study applies it to a real world customer level classification case for implementing customer relationship management. The result shows that the ordinal multiclass SVM model may also be effective for customer level classification.

Comparison of Feature Selection Methods in Support Vector Machines (지지벡터기계의 변수 선택방법 비교)

  • Kim, Kwangsu;Park, Changyi
    • The Korean Journal of Applied Statistics
    • /
    • v.26 no.1
    • /
    • pp.131-139
    • /
    • 2013
  • Support vector machines(SVM) may perform poorly in the presence of noise variables; in addition, it is difficult to identify the importance of each variable in the resulting classifier. A feature selection can improve the interpretability and the accuracy of SVM. Most existing studies concern feature selection in the linear SVM through penalty functions yielding sparse solutions. Note that one usually adopts nonlinear kernels for the accuracy of classification in practice. Hence feature selection is still desirable for nonlinear SVMs. In this paper, we compare the performances of nonlinear feature selection methods such as component selection and smoothing operator(COSSO) and kernel iterative feature extraction(KNIFE) on simulated and real data sets.

Corporate credit rating prediction using support vector machines

  • Lee, Yong-Chan
    • Proceedings of the Korea Inteligent Information System Society Conference
    • /
    • 2005.11a
    • /
    • pp.571-578
    • /
    • 2005
  • Corporate credit rating analysis has drawn a lot of research interests in previous studies, and recent studies have shown that machine learning techniques achieved better performance than traditional statistical ones. This paper applies support vector machines (SVMs) to the corporate credit rating problem in an attempt to suggest a new model with better explanatory power and stability. To serve this purpose, the researcher uses a grid-search technique using 5-fold cross-validation to find out the optimal parameter values of kernel function of SVM. In addition, to evaluate the prediction accuracy of SVM, the researcher compares its performance with those of multiple discriminant analysis (MDA), case-based reasoning (CBR), and three-layer fully connected back-propagation neural networks (BPNs). The experiment results show that SVM outperforms the other methods.

  • PDF

Weighted Support Vector Machines with the SCAD Penalty

  • Jung, Kang-Mo
    • Communications for Statistical Applications and Methods
    • /
    • v.20 no.6
    • /
    • pp.481-490
    • /
    • 2013
  • Classification is an important research area as data can be easily obtained even if the number of predictors becomes huge. The support vector machine(SVM) is widely used to classify a subject into a predetermined group because it gives sound theoretical background and better performance than other methods in many applications. The SVM can be viewed as a penalized method with the hinge loss function and penalty functions. Instead of $L_2$ penalty function Fan and Li (2001) proposed the smoothly clipped absolute deviation(SCAD) satisfying good statistical properties. Despite the ability of SVMs, they have drawbacks of non-robustness when there are outliers in the data. We develop a robust SVM method using a weight function with the SCAD penalty function based on the local quadratic approximation. We compare the performance of the proposed SVM with the SVM using the $L_1$ and $L_2$ penalty functions.

Improving the Generalization Error Bound using Total margin in Support Vector Machines (서포트 벡터 기계에서 TOTAL MARGIN을 이용한 일반화 오차 경계의 개선)

  • Yoon, Min
    • The Korean Journal of Applied Statistics
    • /
    • v.17 no.1
    • /
    • pp.75-88
    • /
    • 2004
  • The Support Vector Machine(SVM) algorithm has paid attention on maximizing the shortest distance between sample points and discrimination hyperplane. This paper suggests the total margin algorithm which considers the distance between all data points and the separating hyperplane. The method extends existing support vector machine algorithm. In addition, this newly proposed method improves the generalization error bound. Numerical experiments show that the total margin algorithm provides good performance, comparing with the previous methods.

A Study on Robustness Improvement of $H_{\infty}$ Control Using SVM (SVM을 이용한 $H_{\infty}$ 제어의 강인성 향상에 관한 연구)

  • Kim, Min-Chan;Yoon, Seong-Sik;Park, Seung-Kyu;Ahn, Ho-Gyun;Kwak, Gun-Pyong;Yoon, Tae-Sung
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.12 no.2
    • /
    • pp.276-281
    • /
    • 2008
  • This paper proposes a new sliding surface which can have the same dynamics of nominal system based on SVM(Support Vector Machines). The conventional sliding mode control can not have the properties of $H_{\infty}$ controller because its sliding surface has lower order dynamics than the original system. The additional states must be used to solve this problem. However, The sliding surface of this paper can have the dynamics of $H_{\infty}$ control system by using support vector machines without defining any additional dynamic state. By using SVM, the property of $H_{\infty}$ control system can be estimated as a relationship between the states. With this relationship, a new sliding surface can be designed and have $H_{\infty}$ control system properties. As a result, in spite of the parameter uncertainty, the proposed controller can have the same dynamic of nominal system controlled by $H_{\infty}$ controller.

Phoneme segmentation and Recognition using Support Vector Machines (Support Vector Machines에 의한 음소 분할 및 인식)

  • Lee, Gwang-Seok;Kim, Deok-Hyun
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2010.05a
    • /
    • pp.981-984
    • /
    • 2010
  • In this paper, we used Support Vector Machines(SVMs) as the learning method, one of Artificial Neural Network, to segregated from the continuous speech into phonemes, an initial, medial, and final sound, and then, performed continuous speech recognition from it. A Decision boundary of phoneme is determined by algorithm with maximum frequency in a short interval. Speech recognition process is performed by Continuous Hidden Markov Model(CHMM), and we compared it with another phoneme segregated from the eye-measurement. From the simulation results, we confirmed that the method, SVMs, we proposed is more effective in an initial sound than Gaussian Mixture Models(GMMs).

  • PDF

Support Vector Bankruptcy Prediction Model with Optimal Choice of RBF Kernel Parameter Values using Grid Search (Support Vector Machine을 이용한 부도예측모형의 개발 -격자탐색을 이용한 커널 함수의 최적 모수 값 선정과 기존 부도예측모형과의 성과 비교-)

  • Min Jae H.;Lee Young-Chan
    • Journal of the Korean Operations Research and Management Science Society
    • /
    • v.30 no.1
    • /
    • pp.55-74
    • /
    • 2005
  • Bankruptcy prediction has drawn a lot of research interests in previous literature, and recent studies have shown that machine learning techniques achieved better performance than traditional statistical ones. This paper employs a relatively new machine learning technique, support vector machines (SVMs). to bankruptcy prediction problem in an attempt to suggest a new model with better explanatory power and stability. To serve this purpose, we use grid search technique using 5-fold cross-validation to find out the optimal values of the parameters of kernel function of SVM. In addition, to evaluate the prediction accuracy of SVM. we compare its performance with multiple discriminant analysis (MDA), logistic regression analysis (Logit), and three-layer fully connected back-propagation neural networks (BPNs). The experiment results show that SVM outperforms the other methods.