• 제목/요약/키워드: kernel methods

Search Result 485, Processing Time 0.028 seconds

A NEW PRIMAL-DUAL INTERIOR POINT METHOD FOR LINEAR OPTIMIZATION

  • Cho, Gyeong-Mi
    • Journal of the Korean Society for Industrial and Applied Mathematics
    • /
    • v.13 no.1
    • /
    • pp.41-53
    • /
    • 2009
  • A primal-dual interior point method(IPM) not only is the most efficient method for a computational point of view but also has polynomial complexity. Most of polynomialtime interior point methods(IPMs) are based on the logarithmic barrier functions. Peng et al.([14, 15]) and Roos et al.([3]-[9]) proposed new variants of IPMs based on kernel functions which are called self-regular and eligible functions, respectively. In this paper we define a new kernel function and propose a new IPM based on this kernel function which has $O(n^{\frac{2}{3}}log\frac{n}{\epsilon})$ and $O(\sqrt{n}log\frac{n}{\epsilon})$ iteration bounds for large-update and small-update methods, respectively.

  • PDF

Data Clustering Method Using a Modified Gaussian Kernel Metric and Kernel PCA

  • Lee, Hansung;Yoo, Jang-Hee;Park, Daihee
    • ETRI Journal
    • /
    • v.36 no.3
    • /
    • pp.333-342
    • /
    • 2014
  • Most hyper-ellipsoidal clustering (HEC) approaches use the Mahalanobis distance as a distance metric. It has been proven that HEC, under this condition, cannot be realized since the cost function of partitional clustering is a constant. We demonstrate that HEC with a modified Gaussian kernel metric can be interpreted as a problem of finding condensed ellipsoidal clusters (with respect to the volumes and densities of the clusters) and propose a practical HEC algorithm that is able to efficiently handle clusters that are ellipsoidal in shape and that are of different size and density. We then try to refine the HEC algorithm by utilizing ellipsoids defined on the kernel feature space to deal with more complex-shaped clusters. The proposed methods lead to a significant improvement in the clustering results over K-means algorithm, fuzzy C-means algorithm, GMM-EM algorithm, and HEC algorithm based on minimum-volume ellipsoids using Mahalanobis distance.

Development of a Surface Modeling Kernel (곡면 모델링 커널 개발)

  • 전차수;구미정;박세형
    • Proceedings of the Korean Society of Precision Engineering Conference
    • /
    • 1996.11a
    • /
    • pp.774-778
    • /
    • 1996
  • Developed in this research is a surface modeling kernel for various CAD/CAM applications. Its internal surface representations are rational parametric polynomials, which are generalizations of nonrational Bezier, Ferguson, Coons and NURBS surface, and are very fast in evaluation. The kernel is designed under the OOP concepts and coded in C++ on PCs. The present implementation of the kernel supports surface construction methods, such as point data interpolation, skinning, sweeping and blending. It also has NURBS conversion routines and offers the IGES and ZES format for geometric information exchange. It includes some geometric processing routines, such as surface/surface intersection, curve/surface intersection, curve projection and so forth. We are continuing to work with the kernel and eventually develop a B-Rep based solid modeler.

  • PDF

Fast Patch-based De-blurring with Directional-oriented Kernel Estimation

  • Min, Kyeongyuk;Chong, Jongwha
    • Journal of IKEEE
    • /
    • v.21 no.1
    • /
    • pp.46-65
    • /
    • 2017
  • This paper proposes a fast patch-based de-blurring algorithm including kernel estimation based on the angle between the edge and the blur direction. For de-blurring, image patches from the most informative edges in the blurry image are used to estimate a kernel with low computational cost. Moreover, the kernels of each patch are estimated based on the correlation between the edge direction and the blur direction. This makes the final kernel more reliable and creates an accurate latent image from the blurry image. The combination of directionally oriented kernel estimation and patch-based de-blurring is faster and more accurate than existing state-of-the art methods. Experimental results using various test images show that the proposed method achieves its objectives: speed and accuracy.

THE STUDY OF FLOOD FREQUENCY ESTIMATES USING CAUCHY VARIABLE KERNEL

  • Moon, Young-Il;Cha, Young-Il;Ashish Sharma
    • Water Engineering Research
    • /
    • v.2 no.1
    • /
    • pp.1-10
    • /
    • 2001
  • The frequency analyses for the precipitation data in Korea were performed. We used daily maximum series, monthly maximum series, and annual series. For nonparametric frequency analyses, variable kernel estimators were used. Nonparametric methods do not require assumptions about the underlying populations from which the data are obtained. Therefore, they are better suited for multimodal distributions with the advantage of not requiring a distributional assumption. In order to compare their performance with parametric distributions, we considered several probability density functions. They are Gamma, Gumbel, Log-normal, Log-Pearson type III, Exponential, Generalized logistic, Generalized Pareto, and Wakeby distributions. The variable kernel estimates are comparable and are in the middle of the range of the parametric estimates. The variable kernel estimates show a very small probability in extrapolation beyond the largest observed data in the sample. However, the log-variable kernel estimates remedied these defects with the log-transformed data.

  • PDF

Weighted Kernel and it's Learning Method for Cancer Diagnosis System (암진단시스템을 위한 Weighted Kernel 및 학습방법)

  • Choi, Gyoo-Seok;Park, Jong-Jin;Jeon, Byoung-Chan;Park, In-Kyu;Ahn, Ihn-Seok;Nguyen, Ha-Nam
    • The Journal of the Institute of Internet, Broadcasting and Communication
    • /
    • v.9 no.2
    • /
    • pp.1-6
    • /
    • 2009
  • One of the most important problems in bioinformatics is how to extract the useful information from a huge amount of data, and make a decision in diagnosis, prognosis, and medical treatment applications. This paper proposes a weighted kernel function for support vector machine and its learning method with a fast convergence and a good classification performance. We defined the weighted kernel function as the weighted sum of a set of different types of basis kernel functions such as neural, radial, and polynomial kernels, which are trained by a learning method based on genetic algorithm. The weights of basis kernel functions in proposed kernel are determined in learning phase and used as the parameters in the decision model in classification phase. The experiments on several clinical datasets such as colon cancer indicate that our weighted kernel function results in higher and more stable classification performance than other kernel functions.

  • PDF

AIT: A method for operating system kernel function call graph generation with a virtualization technique

  • Jiao, Longlong;Luo, Senlin;Liu, Wangtong;Pan, Limin
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.14 no.5
    • /
    • pp.2084-2100
    • /
    • 2020
  • Operating system (OS) kernel function call graphs have been widely used in OS analysis and defense. However, most existing methods and tools for generating function call graphs are designed for application programs, and cannot be used for generating OS kernel function call graphs. This paper proposes a virtualization-based call graph generation method called Acquire in Trap (AIT). When target kernel functions are called, AIT dynamically initiates a system trap with the help of a virtualization technique. It then analyzes and records the calling relationships for trap handling by traversing the kernel stacks and the code space. Our experimental results show that the proposed method is feasible for both Linux and Windows OSs, including 32 and 64-bit versions, with high recall and precision rates. AIT is independent of the source code, compiler and OS kernel architecture, and is a universal method for generating OS kernel function call graphs.

IKPCA-ELM-based Intrusion Detection Method

  • Wang, Hui;Wang, Chengjie;Shen, Zihao;Lin, Dengwei
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.14 no.7
    • /
    • pp.3076-3092
    • /
    • 2020
  • An IKPCA-ELM-based intrusion detection method is developed to address the problem of the low accuracy and slow speed of intrusion detection caused by redundancies and high dimensions of data in the network. First, in order to reduce the effects of uneven sample distribution and sample attribute differences on the extraction of KPCA features, the sample attribute mean and mean square error are introduced into the Gaussian radial basis function and polynomial kernel function respectively, and the two improved kernel functions are combined to construct a hybrid kernel function. Second, an improved particle swarm optimization (IPSO) algorithm is proposed to determine the optimal hybrid kernel function for improved kernel principal component analysis (IKPCA). Finally, IKPCA is conducted to complete feature extraction, and an extreme learning machine (ELM) is applied to classify common attack type detection. The experimental results demonstrate the effectiveness of the constructed hybrid kernel function. Compared with other intrusion detection methods, IKPCA-ELM not only ensures high accuracy rates, but also reduces the detection time and false alarm rate, especially reducing the false alarm rate of small sample attacks.

An edge detection method for gray scale images based on their fuzzy system representation (디지털 영상의 퍼지시스템 표현을 이용한 Edge 검출방법)

  • 문병수;이현철;김장열
    • Journal of the Korean Institute of Intelligent Systems
    • /
    • v.11 no.6
    • /
    • pp.454-458
    • /
    • 2001
  • Based on a fuzzy system representation of gray scale images, we derive and edge detection algorithm whose convolution kernel is different from the known kernels such as those of Robert's Prewitt's or Sobel's gradient. Our fuzzy system representation is an exact representation of the bicubic spline function which represents the gray scale image approximately. Hence the fuzzy system is a continuous function and it provides a natural way to define the gradient and the Laplacian operator. We show that the gradient at grid points can be evaluated by taking the convolution of the image with a 3$\times$3 kernel. We also that our gradient coupled with the approximate value of the continuous function generates an edge detection method which creates edge images clearer than those by other methods. A few examples of applying our methods are included.

  • PDF

Multi-Radial Basis Function SVM Classifier: Design and Analysis

  • Wang, Zheng;Yang, Cheng;Oh, Sung-Kwun;Fu, Zunwei
    • Journal of Electrical Engineering and Technology
    • /
    • v.13 no.6
    • /
    • pp.2511-2520
    • /
    • 2018
  • In this study, Multi-Radial Basis Function Support Vector Machine (Multi-RBF SVM) classifier is introduced based on a composite kernel function. In the proposed multi-RBF support vector machine classifier, the input space is divided into several local subsets considered for extremely nonlinear classification tasks. Each local subset is expressed as nonlinear classification subspace and mapped into feature space by using kernel function. The composite kernel function employs the dual RBF structure. By capturing the nonlinear distribution knowledge of local subsets, the training data is mapped into higher feature space, then Multi-SVM classifier is realized by using the composite kernel function through optimization procedure similar to conventional SVM classifier. The original training data set is partitioned by using some unsupervised learning methods such as clustering methods. In this study, three types of clustering method are considered such as Affinity propagation (AP), Hard C-Mean (HCM) and Iterative Self-Organizing Data Analysis Technique Algorithm (ISODATA). Experimental results on benchmark machine learning datasets show that the proposed method improves the classification performance efficiently.