• Title/Summary/Keyword: Kernel Method

Search Result 993, Processing Time 0.023 seconds

Analysis of Bulk Metal Forming Process by Reproducing Kernel Particle Method (재생커널입자법을 이용한 체적성형공정의 해석)

  • Han, Kyu-Taek
    • Journal of the Korean Society of Manufacturing Process Engineers
    • /
    • v.8 no.3
    • /
    • pp.21-26
    • /
    • 2009
  • The finite element analysis of metal forming processes often fails because of severe mesh distortion at large deformation. As the concept of meshless methods, only nodal point data are used for modeling and solving. As the main feature of these methods, the domain of the problem is represented by a set of nodes, and a finite element mesh is unnecessary. This computational methods reduces time-consuming model generation and refinement effort. It provides a higher rate of convergence than the conventional finite element methods. The displacement shape functions are constructed by the reproducing kernel approximation that satisfies consistency conditions. In this research, A meshless method approach based on the reproducing kernel particle method (RKPM) is applied with metal forming analysis. Numerical examples are analyzed to verify the performance of meshless method for metal forming analysis.

  • PDF

Adaptive kernel method for evaluating structural system reliability

  • Wang, G.S.;Ang, A.H.S.;Lee, J.C.
    • Structural Engineering and Mechanics
    • /
    • v.5 no.2
    • /
    • pp.115-126
    • /
    • 1997
  • Importance sampling methods have been developed with the aim of reducing the computational costs inherent in Monte Carlo methods. This study proposes a new algorithm called the adaptive kernel method which combines and modifies some of the concepts from adaptive sampling and the simple kernel method to evaluate the structural reliability of time variant problems. The essence of the resulting algorithm is to select an appropriate starting point from which the importance sampling density can be generated efficiently. Numerical results show that the method is unbiased and substantially increases the efficiency over other methods.

Kernel Hardening by Recovering Kernel Stack Frame in Linux Operating System (리눅스 운영체제에서 커널 스택의 복구를 통한 커널 하드닝)

  • Jang Seung-Ju
    • The KIPS Transactions:PartA
    • /
    • v.13A no.3 s.100
    • /
    • pp.199-204
    • /
    • 2006
  • The kernel hardening function is necessary in terms of kernel stability to reduce the system error or panic due to the kernel code error that is made by program developer. But, the traditional kernel hardening method is difficult to implement and consuming high cost. The suggested kernel hardening function that makes high availability system by changing the panic() function of inside kernel code guarantees normal system operation by recovering the incorrect address of the kernel stack frame. We experimented the kernel hardening function at the network module of the Linux by forcing panic code and confirmed the proposed design mechanism of kernel hardening is working well by this experiment.

On Predicting with Kernel Ridge Regression

  • Hwang, Chang-Ha
    • Journal of the Korean Data and Information Science Society
    • /
    • v.14 no.1
    • /
    • pp.103-111
    • /
    • 2003
  • Kernel machines are used widely in real-world regression tasks. Kernel ridge regressions(KRR) and support vector machines(SVM) are typical kernel machines. Here, we focus on two types of KRR. One is inductive KRR. The other is transductive KRR. In this paper, we study how differently they work in the interpolation and extrapolation areas. Furthermore, we study prediction interval estimation method for KRR. This turns out to be a reliable and practical measure of prediction interval and is essential in real-world tasks.

  • PDF

Arrow Diagrams for Kernel Principal Component Analysis

  • Huh, Myung-Hoe
    • Communications for Statistical Applications and Methods
    • /
    • v.20 no.3
    • /
    • pp.175-184
    • /
    • 2013
  • Kernel principal component analysis(PCA) maps observations in nonlinear feature space to a reduced dimensional plane of principal components. We do not need to specify the feature space explicitly because the procedure uses the kernel trick. In this paper, we propose a graphical scheme to represent variables in the kernel principal component analysis. In addition, we propose an index for individual variables to measure the importance in the principal component plane.

Variable selection in the kernel Cox regression

  • Shim, Joo-Yong
    • Journal of the Korean Data and Information Science Society
    • /
    • v.22 no.4
    • /
    • pp.795-801
    • /
    • 2011
  • In machine learning and statistics it is often the case that some variables are not important, while some variables are more important than others. We propose a novel algorithm for selecting such relevant variables in the kernel Cox regression. We employ the weighted version of ANOVA decomposition kernels to choose optimal subset of relevant variables in the kernel Cox regression. Experimental results are then presented which indicate the performance of the proposed method.

Speaker Verification Using SVM Kernel with GMM-Supervector Based on the Mahalanobis Distance (Mahalanobis 거리측정 방법 기반의 GMM-Supervector SVM 커널을 이용한 화자인증 방법)

  • Kim, Hyoung-Gook;Shin, Dong
    • The Journal of the Acoustical Society of Korea
    • /
    • v.29 no.3
    • /
    • pp.216-221
    • /
    • 2010
  • In this paper, we propose speaker verification method using Support Vector Machine (SVM) kernel with Gaussian Mixture Model (GMM)-supervector based on the Mahalanobis distance. The proposed GMM-supervector SVM kernel method is combined GMM with SVM. The GMM-supervectors are generated by GMM parameters of speaker and other speaker utterances. A speaker verification threshold of GMM-supervectors is decided by SVM kernel based on Mahalanobis distance to improve speaker verification accuracy. The experimental results for text-independent speaker verification using 20 speakers demonstrates the performance of the proposed method compared to GMM, SVM, GMM-supervector SVM kernel based on Kullback-Leibler (KL) divergence, and GMM-supervector SVM kernel based on Bhattacharyya distance.

Ensemble approach for improving prediction in kernel regression and classification

  • Han, Sunwoo;Hwang, Seongyun;Lee, Seokho
    • Communications for Statistical Applications and Methods
    • /
    • v.23 no.4
    • /
    • pp.355-362
    • /
    • 2016
  • Ensemble methods often help increase prediction ability in various predictive models by combining multiple weak learners and reducing the variability of the final predictive model. In this work, we demonstrate that ensemble methods also enhance the accuracy of prediction under kernel ridge regression and kernel logistic regression classification. Here we apply bagging and random forests to two kernel-based predictive models; and present the procedure of how bagging and random forests can be embedded in kernel-based predictive models. Our proposals are tested under numerous synthetic and real datasets; subsequently, they are compared with plain kernel-based predictive models and their subsampling approach. Numerical studies demonstrate that ensemble approach outperforms plain kernel-based predictive models.

An Overview of Unsupervised and Semi-Supervised Fuzzy Kernel Clustering

  • Frigui, Hichem;Bchir, Ouiem;Baili, Naouel
    • International Journal of Fuzzy Logic and Intelligent Systems
    • /
    • v.13 no.4
    • /
    • pp.254-268
    • /
    • 2013
  • For real-world clustering tasks, the input data is typically not easily separable due to the highly complex data structure or when clusters vary in size, density and shape. Kernel-based clustering has proven to be an effective approach to partition such data. In this paper, we provide an overview of several fuzzy kernel clustering algorithms. We focus on methods that optimize an fuzzy C-mean-type objective function. We highlight the advantages and disadvantages of each method. In addition to the completely unsupervised algorithms, we also provide an overview of some semi-supervised fuzzy kernel clustering algorithms. These algorithms use partial supervision information to guide the optimization process and avoid local minima. We also provide an overview of the different approaches that have been used to extend kernel clustering to handle very large data sets.

TIME DISCRETIZATION WITH SPATIAL COLLOCATION METHOD FOR A PARABOLIC INTEGRO-DIFFERENTIAL EQUATION WITH A WEAKLY SINGULAR KERNEL

  • Kim Chang-Ho
    • The Pure and Applied Mathematics
    • /
    • v.13 no.1 s.31
    • /
    • pp.19-38
    • /
    • 2006
  • We analyze the spectral collocation approximation for a parabolic partial integrodifferential equations(PIDE) with a weakly singular kernel. The space discretization is based on the spectral collocation method and the time discretization is based on Crank-Nicolson scheme with a graded mesh. We obtain the stability and second order convergence result for fully discrete scheme.

  • PDF