• Title/Summary/Keyword: Kernel method

Search Result 993, Processing Time 0.028 seconds

Function Approximation Based on a Network with Kernel Functions of Bounds and Locality : an Approach of Non-Parametric Estimation

  • Kil, Rhee-M.
    • ETRI Journal
    • /
    • v.15 no.2
    • /
    • pp.35-51
    • /
    • 1993
  • This paper presents function approximation based on nonparametric estimation. As an estimation model of function approximation, a three layered network composed of input, hidden and output layers is considered. The input and output layers have linear activation units while the hidden layer has nonlinear activation units or kernel functions which have the characteristics of bounds and locality. Using this type of network, a many-to-one function is synthesized over the domain of the input space by a number of kernel functions. In this network, we have to estimate the necessary number of kernel functions as well as the parameters associated with kernel functions. For this purpose, a new method of parameter estimation in which linear learning rule is applied between hidden and output layers while nonlinear (piecewise-linear) learning rule is applied between input and hidden layers, is considered. The linear learning rule updates the output weights between hidden and output layers based on the Linear Minimization of Mean Square Error (LMMSE) sense in the space of kernel functions while the nonlinear learning rule updates the parameters of kernel functions based on the gradient of the actual output of network with respect to the parameters (especially, the shape) of kernel functions. This approach of parameter adaptation provides near optimal values of the parameters associated with kernel functions in the sense of minimizing mean square error. As a result, the suggested nonparametric estimation provides an efficient way of function approximation from the view point of the number of kernel functions as well as learning speed.

  • PDF

Fuzzy Kernel K-Nearest Neighbor Algorithm for Image Segmentation (영상 분할을 위한 퍼지 커널 K-nearest neighbor 알고리즘)

  • Choi Byung-In;Rhee Chung-Hoon
    • Journal of the Korean Institute of Intelligent Systems
    • /
    • v.15 no.7
    • /
    • pp.828-833
    • /
    • 2005
  • Kernel methods have shown to improve the performance of conventional linear classification algorithms for complex distributed data sets, as mapping the data in input space into a higher dimensional feature space(7). In this paper, we propose a fuzzy kernel K-nearest neighbor(fuzzy kernel K-NN) algorithm, which applies the distance measure in feature space based on kernel functions to the fuzzy K-nearest neighbor(fuzzy K-NN) algorithm. In doing so, the proposed algorithm can enhance the Performance of the conventional algorithm, by choosing an appropriate kernel function. Results on several data sets and segmentation results for real images are given to show the validity of our proposed algorithm.

Spare Representation Learning of Kernel Space Using the Kernel Relaxation Procedure (커널 이완 절차에 의한 커널 공간의 저밀도 표현 학습)

  • 류재홍;정종철
    • Journal of the Korean Institute of Intelligent Systems
    • /
    • v.11 no.9
    • /
    • pp.817-821
    • /
    • 2001
  • In this paper, a new learning methodology for kernel methods that results in a sparse representation of kernel space from the training patterns for classification problems is suggested. Among the traditional algorithms of linear discriminant function, this paper shows that the relaxation procedure can obtain the maximum margin separating hyperplane of linearly separable pattern classification problem as SVM(Support Vector Machine) classifier does. The original relaxation method gives only the necessary condition of SV patterns. We suggest the sufficient condition to identify the SV patterns in the learning epoches. For sequential learning of kernel methods, extended SVM and kernel discriminant function are defined. Systematic derivation of learning algorithm is introduced. Experiment results show the new methods have the higher or equivalent performance compared to the conventional approach.

  • PDF

Sparse Representation Learning of Kernel Space Using the Kernel Relaxation Procedure (커널 이완절차에 의한 커널 공간의 저밀도 표현 학습)

  • 류재홍;정종철
    • Proceedings of the Korean Institute of Intelligent Systems Conference
    • /
    • 2001.12a
    • /
    • pp.60-64
    • /
    • 2001
  • In this paper, a new learning methodology for Kernel Methods is suggested that results in a sparse representation of kernel space from the training patterns for classification problems. Among the traditional algorithms of linear discriminant function(perceptron, relaxation, LMS(least mean squared), pseudoinverse), this paper shows that the relaxation procedure can obtain the maximum margin separating hyperplane of linearly separable pattern classification problem as SVM(Support Vector Machine) classifier does. The original relaxation method gives only the necessary condition of SV patterns. We suggest the sufficient condition to identify the SV patterns in the learning epochs. Experiment results show the new methods have the higher or equivalent performance compared to the conventional approach.

  • PDF

Directional Information Processing Using Optical Rotating Kernel Operations (광회전 커널 오퍼레이션을 이용하는 방향성 정보 처리)

  • Yim Kul Lee
    • Journal of the Korean Institute of Telematics and Electronics B
    • /
    • v.30B no.2
    • /
    • pp.78-86
    • /
    • 1993
  • A nonlinear method for directional information processing is introduced, along with an application of directional feature enhancement. In this method, an input is convolved with a 2-D ong, norrow kernel, which is rotated through 360 degree, continuously or discretely in a large number of steps. An output is given by some function of the convolution results. Linear features that are aligned with the kernel are enhanced, otherwise, removed or suppressed. The method presented is insensitive to variation in the dimension of linear features to be processed and preserves a good enhancement capability even for an image characterized by low contrast and spatially varying brightness in noisy backgroung. Effects of the kernel legnth and width on the performance are discussed. A possible hybrid optical-electronic implementation is also discussed.

  • PDF

AN ELIGIBLE PRIMAL-DUAL INTERIOR-POINT METHOD FOR LINEAR OPTIMIZATION

  • Cho, Gyeong-Mi;Lee, Yong-Hoon
    • East Asian mathematical journal
    • /
    • v.29 no.3
    • /
    • pp.279-292
    • /
    • 2013
  • It is well known that each kernel function defines a primal-dual interior-point method(IPM). Most of polynomial-time interior-point algorithms for linear optimization(LO) are based on the logarithmic kernel function([2, 11]). In this paper we define a new eligible kernel function and propose a new search direction and proximity function based on this function for LO problems. We show that the new algorithm has ${\mathcal{O}}((log\;p){\sqrt{n}}\;log\;n\;log\;{\frac{n}{\epsilon}})$ and ${\mathcal{O}}((q\;log\;p)^{\frac{3}{2}}{\sqrt{n}}\;log\;{\frac{n}{\epsilon}})$ iteration bound for large- and small-update methods, respectively. These are currently the best known complexity results.

Estimating Variance Function with Kernel Machine

  • Kim, Jong-Tae;Hwang, Chang-Ha;Park, Hye-Jung;Shim, Joo-Yong
    • Communications for Statistical Applications and Methods
    • /
    • v.16 no.2
    • /
    • pp.383-388
    • /
    • 2009
  • In this paper we propose a variance function estimation method based on kernel trick for replicated data or data consisted of sample variances. Newton-Raphson method is used to obtain associated parameter vector. Furthermore, the generalized approximate cross validation function is introduced to select the hyper-parameters which affect the performance of the proposed variance function estimation method. Experimental results are then presented which illustrate the performance of the proposed procedure.

Estimating multiplicative competitive interaction model using kernel machine technique

  • Shim, Joo-Yong;Kim, Mal-Suk;Park, Hye-Jung
    • Journal of the Korean Data and Information Science Society
    • /
    • v.23 no.4
    • /
    • pp.825-832
    • /
    • 2012
  • We propose a novel way of forecasting the market shares of several brands simultaneously in a multiplicative competitive interaction model, which uses kernel regression technique incorporated with kernel machine technique applied in support vector machines and other machine learning techniques. Traditionally, the estimations of the market share attraction model are performed via a maximum likelihood estimation procedure under the assumption that the data are drawn from a normal distribution. The proposed method is shown to be a good candidate for forecasting method of the market share attraction model when normal distribution is not assumed. We apply the proposed method to forecast the market shares of 4 Korean car brands simultaneously and represent better performances than maximum likelihood estimation procedure.

Numerical Solution For Fredholm Integral Equation With Hilbert Kernel

  • Abdou, Mohamed Abdella Ahmed;Hendi, Fathea Ahmed
    • Journal of the Korean Society for Industrial and Applied Mathematics
    • /
    • v.9 no.1
    • /
    • pp.111-123
    • /
    • 2005
  • Here, the Fredholm integral equation with Hilbert kernel is solved numerically, using two different methods. Also the error, in each case, is estimated.

  • PDF