Kernel PCA
Assume that we are dealing with centred data
The covariance matrix then takes from
Then we have to find eigenvalues and nonzero eigenvectors v in Hilbert space satisfying:
All solutions v with lie in the span of
, due to the fact that
(
is a inner product between two vectors, it is a scalar. So we can reorder it to the front. )
We can get 2 useful consequence based on the former equation, the first is:
If we choose a point and multiply it on both side. We can get
The second consequence is that the eigenvector can be written as a linear combination of points in Hilbert space:
(
)
Replace consequence 2 into 1:
We can get (43)
There are two properties of inner product that we should know is that:
(
is a constant)
(
is a constant)
Have known these two properties, we can write equation (43) as:
This can be rewritten, for all j=1,…,n, as:
; (44)
; (45)
. (46)
The two underprices implies that:
(This is the definition of matrix multiplication: the jth row of K multiply with the ithcolumn of K is the ij entry of the result matrix.)
Kernel PCA as an eigenvector problem
Equation (46) can be rewritten as:
(47)
where denotes the column vector with entries
To find the solution of Equation (47), we solve the problem
which we obtain by multiplying (47) by from the left.
Normalizing the coefficients
We require that the eigenvectors to have unit length, that is
for all k=1,…,r.
That means that
As eigenvectors of , the
have unit form.
Therefor we have to rescale them by to enforce that their norm is
.