What is SVM, and how does it relate to SVC?
H
How do we use basis expansion in SVC optimisations to enlarge the space of features by including squared terms?
If we use a basis expansion of cubic polynomials on (x(1), x(2)</sup), how does the SVC optimisation problem change?
What is the problem with specifying a large basis?
What method do we use to efficiently enlarge our feature space to accommodate a nonlinear boundary between classes?
What is the key idea about SVC we will be using?
How can both the MMC and the SVC be represented as their inner product?
What is the informal idea behind using the kernel and its association with the inner product?
How do we conduct SVM using a kernel?
Can we take the original ridge regression optimisation problem and represent it in the form of an inner product?
Thus, can we apply the kernel method under the ridge regression?
What is the summary of the intuition behind using the kernel?
So what question do we have when using the kernel?
What is the mathematical setup, features and variables used in the kernel method?
(6 bullet points)
How do we define the (real) vector space?
What is an inner product space? What properties does it have to satisfy?
What is the notion of “length” given by in this space?
What is a Hilbert space?
What is the formal definition of a Kernel?
What is the formal definition of a (Positive semi-definite kernel)?
What is the definition of a positive semi-definite matrix?
kernel <–> positive-semi definite kernel
Given k1 and k2 are kernels, what else is a kernel?
What is the:
What are the three equivalent definitions of a Kernel: Moore-Aronszajn Theorem?
How do we build towards the reproducing kernel Hilbert space ( RKHS)?
What is the formal definition of a Reproducing Kernel Hilbert Space (RKHS)?
What is the Representer Theorem?
With a kernel, what does the SVM classifier for a new observation depend on?
What are four popular kernels?
How do SVM work with a radial kernel?
What is the advantage of kernel methods?