Solved – What are the advantages of Multiple Kernel Learning (MKL) methods

kernel trickmachine learningsvm

Multiple Kernel Learning methods aim to construct a kernel model where the kernel is a linear combination of fixed base kernels. Learning the kernel then consists of learning the weighting coefficients for each base kernel, rather than optimising the kernel parameters of a single kernel.

The disadvantages of multiple kernel learning seem to be that they are less interpretable and computationally expensive (as to evaluate the model output you need to evaluate all of the base kernels). So if similar performance can be achieved by simply optimising a single kernel, what are the advantages of MKL?

Best Answer

There are two advantages (or rather two use-cases):

  1. For every application of SVMs, a user has to choose which kernel to use and sometimes even have to design their own kernel matrices. Is it possible to alleviate choosing kernels or specialized kernel designs? MKL was a step towards that.

  2. The second case IMHO is by far a more compelling case. Consider that your data input is a video data + cc. The feature representation of each video consists of video features, audio features and text features. Such a data is known as multi-modal data. Each set of these features may require a different notion of similarity (a different kernel). Instead of building a specialized kernel for such applications, is it possible to just define kernel for each of these modes and linearly combine them?

Related Question