Data-dependent Kernel Support Vector Machine classifiers in Reproducing Kernel Hilbert Space

Abstract Support Vector Machine (SVM) provides a linear classifier for binary classification problems. Complex decision boundaries in the input feature space are handled by nonlinear kernels to the SVM. Theories in Reproducing Kernel Hilbert Spaces (RKHS) state that, given a kernel \(\mathcal{K}\) and a set of \(M\) given data \((x_i,y_i)\), for \(i=1,\cdots,M\), a SVM classifier function can be written as \(f(x) =\alpha_0 + \sum_{i=1}^M \alpha_i \mathcal{K}(x, x_i)\) for some coefficients \(\alpha_i\)s. Also, applying conformal transforms to a positive definite kernel produces another positive definite kernel which are in more complexity. Hence, in case that well-known kernels fail given the current training data, a new kernel can be tried by optimizing the coefficients of a conformal kernel in the way to maximize the ratio “(Between-class error)/(Within-class error)” of the training data. Here, data-dependent kernel SVM is applied to an application of classifying tumor/tumor-free organs from gene expression data and compared its classification performance with other well-known kernels.

View file




Enjoy Reading This Article?

Here are some more articles you might like to read next:

  • Wasserstein Proximals Stabilize Training of Generative Models and Learn Manifolds
  • Lipschitz-Regularized Gradient Flows and Latent Generative Particles
  • Lipschitz Regularized Gradient Flows and Latent Generative Particles
  • Sample generation from unknown distributions - Particle Descent Algorithm induced by (f, Γ)-gradient flow
  • Python ODE solving tutorial