• Title/Summary/Keyword: support vector

Search Result 2,315, Processing Time 0.03 seconds

Modifying linearly non-separable support vector machine binary classifier to account for the centroid mean vector

  • Mubarak Al-Shukeili;Ronald Wesonga
    • Communications for Statistical Applications and Methods
    • /
    • v.30 no.3
    • /
    • pp.245-258
    • /
    • 2023
  • This study proposes a modification to the objective function of the support vector machine for the linearly non-separable case of a binary classifier yi ∈ {-1, 1}. The modification takes into account the position of each data item xi from its corresponding class centroid. The resulting optimization function involves the centroid mean vector, and the spread of data besides the support vectors, which should be minimized by the choice of hyper-plane β. Theoretical assumptions have been tested to derive an optimal separable hyperplane that yields the minimal misclassification rate. The proposed method has been evaluated using simulation studies and real-life COVID-19 patient outcome hospitalization data. Results show that the proposed method performs better than the classical linear SVM classifier as the sample size increases and is preferred in the presence of correlations among predictors as well as among extreme values.

Generalized Support Vector Quantile Regression (일반화 서포트벡터 분위수회귀에 대한 연구)

  • Lee, Dongju;Choi, Sujin
    • Journal of Korean Society of Industrial and Systems Engineering
    • /
    • v.43 no.4
    • /
    • pp.107-115
    • /
    • 2020
  • Support vector regression (SVR) is devised to solve the regression problem by utilizing the excellent predictive power of Support Vector Machine. In particular, the ⲉ-insensitive loss function, which is a loss function often used in SVR, is a function thatdoes not generate penalties if the difference between the actual value and the estimated regression curve is within ⲉ. In most studies, the ⲉ-insensitive loss function is used symmetrically, and it is of interest to determine the value of ⲉ. In SVQR (Support Vector Quantile Regression), the asymmetry of the width of ⲉ and the slope of the penalty was controlled using the parameter p. However, the slope of the penalty is fixed according to the p value that determines the asymmetry of ⲉ. In this study, a new ε-insensitive loss function with p1 and p2 parameters was proposed. A new asymmetric SVR called GSVQR (Generalized Support Vector Quantile Regression) based on the new ε-insensitive loss function can control the asymmetry of the width of ⲉ and the slope of the penalty using the parameters p1 and p2, respectively. Moreover, the figures show that the asymmetry of the width of ⲉ and the slope of the penalty is controlled. Finally, through an experiment on a function, the accuracy of the existing symmetric Soft Margin, asymmetric SVQR, and asymmetric GSVQR was examined, and the characteristics of each were shown through figures.

Multi-pattern Classification Using Kernel Bagging-based Import Vector Machine (커널 Bagging기반의 Import Vector Machine을 이용한 다중 패턴 분류)

  • 최준혁;김대수;임기욱
    • Proceedings of the Korean Institute of Intelligent Systems Conference
    • /
    • 2002.12a
    • /
    • pp.275-278
    • /
    • 2002
  • Vapnik이 제안한 Support Vector Machine은 두 개의 부류를 갖는 데이터에 대한 분류에는 매우 좋은 성능을 보인다는 점은 이미 잘 알려져 있다. 하지만 부류의 개수가 3개 이상인 다중 패턴을 갖는 데이터에 대한 분류에는 SVM을 적용하기가 쉽지 않다. Support Vector Machine의 이러한 문제점을 해결하기 위하여 Zhu는 3개 이상의 부류를 갖는 데이터의 패턴 분류를 위하여 Import Vector Machine을 제안하였다. 이 모형은 Support Vector Machine을 이용하여 해결하기 어려운 다중 패턴 분류를 가능케 한다. Import Vector Machine은 커널 로지스틱 기반의 함수만을 사용하지만 본 논문에서는 다수의 커널 함수를 적용하여 가장 성능이 우수한 커널 함수를 찾아내어 최종 분류를 수행하게되는 bagging 기법을 적용하였다 제안하는 방법이 기존의 방법에 비해, 더욱 정확한 분류를 수행함을 실험 결과를 통해 확인한다.

Improving Learning Performance of Support Vector Machine using the Kernel Relaxation and the Dynamic Momentum (Kernel Relaxation과 동적 모멘트를 조합한 Support Vector Machine의 학습 성능 향상)

  • Kim, Eun-Mi;Lee, Bae-Ho
    • The KIPS Transactions:PartB
    • /
    • v.9B no.6
    • /
    • pp.735-744
    • /
    • 2002
  • This paper proposes learning performance improvement of support vector machine using the kernel relaxation and the dynamic momentum. The dynamic momentum is reflected to different momentum according to current state. While static momentum is equally influenced on the whole, the proposed dynamic momentum algorithm can control to the convergence rate and performance according to the change of the dynamic momentum by training. The proposed algorithm has been applied to the kernel relaxation as the new sequential learning method of support vector machine presented recently. The proposed algorithm has been applied to the SONAR data which is used to the standard classification problems for evaluating neural network. The simulation results of proposed algorithm have better the convergence rate and performance than those using kernel relaxation and static momentum, respectively.

Learning and Performance Comparison of Multi-class Classification Problems based on Support Vector Machine (지지벡터기계를 이용한 다중 분류 문제의 학습과 성능 비교)

  • Hwang, Doo-Sung
    • Journal of Korea Multimedia Society
    • /
    • v.11 no.7
    • /
    • pp.1035-1042
    • /
    • 2008
  • The support vector machine, as a binary classifier, is known to surpass the other classifiers only in binary classification problems through the various experiments. Even though its theory is based on the maximal margin classifier, the support vector machine approach cannot be easily extended to the multi-classification problems. In this paper, we review the extension techniques of the support vector machine toward the multi-classification and do the performance comparison. Depending on the data decomposition of the training data, the support vector machine is easily adapted for a multi-classification problem without modifying the intrinsic characteristics of the binary classifier. The performance is evaluated on a collection of the benchmark data sets and compared according to the selected teaming strategies, the training time, and the results of the neural network with the backpropagation teaming. The experiments suggest that the support vector machine is applicable and effective in the general multi-class classification problems when compared to the results of the neural network.

  • PDF

A Study on Hierarchical Distributed Intrusion Detection for Secure Home Networks Service (안전한 홈네트워크 서비스를 위한 계층적 분산 침입탐지에 관한 연구)

  • Yu, Jae-Hak;Choi, Sung-Back;Yang, Sung-Hyun;Park, Dai-Hee;Chung, Yong-Wha
    • Journal of the Korea Institute of Information Security & Cryptology
    • /
    • v.18 no.1
    • /
    • pp.49-57
    • /
    • 2008
  • In this paper, we propose a novel hierarchical distributed intrusion detection system, named HNHDIDS(Home Network Hierarchical Distributed Intrusion Detection System), which is not only based on the structure of distributed intrusion detection system, but also fully consider the environment of secure home networks service. The proposed system is hierarchically composed of the one-class support vector machine(support vector data description) and local agents, in which it is designed for optimizing for the environment of secure home networks service. We support our findings with computer experiments and analysis.

The Use of MSVM and HMM for Sentence Alignment

  • Fattah, Mohamed Abdel
    • Journal of Information Processing Systems
    • /
    • v.8 no.2
    • /
    • pp.301-314
    • /
    • 2012
  • In this paper, two new approaches to align English-Arabic sentences in bilingual parallel corpora based on the Multi-Class Support Vector Machine (MSVM) and the Hidden Markov Model (HMM) classifiers are presented. A feature vector is extracted from the text pair that is under consideration. This vector contains text features such as length, punctuation score, and cognate score values. A set of manually prepared training data was assigned to train the Multi-Class Support Vector Machine and Hidden Markov Model. Another set of data was used for testing. The results of the MSVM and HMM outperform the results of the length based approach. Moreover these new approaches are valid for any language pairs and are quite flexible since the feature vector may contain less, more, or different features, such as a lexical matching feature and Hanzi characters in Japanese-Chinese texts, than the ones used in the current research.

Face Detection Using Support Vector Domain Description in Color Images (컬러 영상에서 Support Vector Domain Description을 이용한 얼굴 검출)

  • Seo Jin;Ko Hanseok
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.42 no.1
    • /
    • pp.25-31
    • /
    • 2005
  • In this paper, we present a face detection system using the Support Vector Domain Description (SVDD) in color images. Conventional face detection algorithms require a training procedure using both face and non-face images. In SVDD however we employ only face images for training. We can detect faces in color images from the radius and center pairs of SVDD. We also use Entropic Threshold for extracting the facial feature and sliding window for improved performance while saving processing time. The experimental results indicate the effectiveness and efficiency of the proposed algorithm compared to conventional PCA (Principal Component Analysis)-based methods.

Parameter Tuning in Support Vector Regression for Large Scale Problems (대용량 자료에 대한 서포트 벡터 회귀에서 모수조절)

  • Ryu, Jee-Youl;Kwak, Minjung;Yoon, Min
    • Journal of the Korean Institute of Intelligent Systems
    • /
    • v.25 no.1
    • /
    • pp.15-21
    • /
    • 2015
  • In support vector machine, the values of parameters included in kernels affect strongly generalization ability. It is often difficult to determine appropriate values of those parameters in advance. It has been observed through our studies that the burden for deciding the values of those parameters in support vector regression can be reduced by utilizing ensemble learning. However, the straightforward application of the method to large scale problems is too time consuming. In this paper, we propose a method in which the original data set is decomposed into a certain number of sub data set in order to reduce the burden for parameter tuning in support vector regression with large scale data sets and imbalanced data set, particularly.