机器学习-白板推导 P5_4
PCA最小重构代价
X
=
[
x
1
x
2
.
.
.
x
N
]
T
=
[
x
1
T
x
2
T
⋮
x
N
T
]
=
[
x
11
x
12
.
.
.
x
1
p
x
21
x
22
.
.
.
x
2
p
⋮
⋮
⋱
⋮
x
N
1
x
N
2
.
.
.
x
N
p
]
N
∗
p
X=\begin{bmatrix} x_1 & x_2 &...& x_N \end{bmatrix}^T=\begin{bmatrix} x_1^T \\ x_2^T \\\vdots\\ x_N^T \end{bmatrix} = \begin{bmatrix} x_{11} & x_{12} &...& x_{1p} \\ x_{21} & x_{22} &...& x_{2p} \\ \vdots & \vdots & \ddots & \vdots \\ x_{N1} & x_{N2} &...& x_{Np} \\ \end{bmatrix}_{N*p}
X=[x1x2...xN]T=⎣⎢⎢⎢⎡x1Tx2T⋮xNT⎦⎥⎥⎥⎤=⎣⎢⎢⎢⎡x11x21⋮xN1x12x22⋮xN2......⋱...x1px2p⋮xNp⎦⎥⎥⎥⎤N∗p
1
N
=
[
1
1
⋮
1
]
1_N= \begin{bmatrix} 1 \\ 1 \\\vdots\\ 1 \end{bmatrix}
1N=⎣⎢⎢⎢⎡11⋮1⎦⎥⎥⎥⎤
M
e
a
n
:
X
‾
=
1
N
∑
i
=
1
N
x
i
=
1
N
X
T
1
N
Mean:\overline{X}=\frac{1}{N}\sum_{i=1}^N x_i=\frac{1}{N}X^T1_N
Mean:X=N1∑i=1Nxi=N1XT1N
C o v a r i a n c e : S = 1 N ∑ i = 1 N ( x i − X ‾ ) ( x i − X ‾ ) T = 1 N X T H X Covariance:S = \frac{1}{N}\sum_{i=1}^N (x_i- \overline{X})(x_i- \overline{X})^T= \frac{1}{N}X^THX Covariance:S=N1∑i=1N(xi−X)(xi−X)T=N1XTHX
x
i
x_i
xi在
u
k
u_k
uk方向的投影值为
x
i
T
u
k
x_i^Tu_k
xiTuk,投影值和单位向量的乘机为
(
x
i
T
u
k
)
u
k
(x_i^Tu_k)u_k
(xiTuk)uk
所以
x i = ∑ k = 1 p ( x i T u k ) u k x_i=\sum_{k=1}^p(x_i^Tu_k)u_k xi=∑k=1p(xiTuk)uk
用PCA进行降维,降维后的维度为 q < p q<p q<p
x i ^ = ∑ k = 1 q ( x i T u k ) u k \hat{x_i}=\sum_{k=1}^q(x_i^Tu_k)u_k xi^=∑k=1q(xiTuk)uk
重构代价为:
J
=
1
N
∑
i
=
1
N
∣
∣
x
i
−
x
i
^
∣
∣
2
=
1
N
∑
i
=
1
N
∣
∣
∑
k
=
q
+
1
p
(
x
i
T
u
k
)
u
k
∣
∣
2
=
1
N
∑
i
=
1
N
∑
k
=
q
+
1
p
(
x
i
T
u
k
)
2
\begin{aligned} J &= \frac{1}{N} \sum_{i=1}^N ||x_i - \hat{x_i}||^2 \\ &= \frac{1}{N} \sum_{i=1}^N ||\sum_{k=q+1}^p(x_i^Tu_k)u_k||^2 \\ &= \frac{1}{N}\sum_{i=1}^N \sum_{k=q+1}^p(x_i^Tu_k)^2 \end{aligned}
J=N1i=1∑N∣∣xi−xi^∣∣2=N1i=1∑N∣∣k=q+1∑p(xiTuk)uk∣∣2=N1i=1∑Nk=q+1∑p(xiTuk)2
x
i
x_i
xi进行中心化:
J
=
1
N
∑
i
=
1
N
∑
k
=
q
+
1
p
(
(
x
i
−
X
‾
)
T
u
k
)
2
=
∑
k
=
q
+
1
p
∑
i
=
1
N
1
N
(
(
x
i
−
X
‾
)
T
u
k
)
2
=
∑
k
=
q
+
1
p
u
k
T
S
u
k
s
.
t
.
u
k
T
u
k
=
1
\begin{aligned} J &= \frac{1}{N}\sum_{i=1}^N \sum_{k=q+1}^p((x_i-\overline{X})^Tu_k)^2 \\ &= \sum_{k=q+1}^p \sum_{i=1}^N \frac{1}{N} ((x_i-\overline{X})^Tu_k)^2 \\ &=\sum_{k=q+1}^p u_k^TSu_k \\ &s.t. \quad u_k^Tu_k=1 \end{aligned}
J=N1i=1∑Nk=q+1∑p((xi−X)Tuk)2=k=q+1∑pi=1∑NN1((xi−X)Tuk)2=k=q+1∑pukTSuks.t.ukTuk=1
最小重构代价:
J
=
1
N
∑
i
=
1
N
∣
∣
x
i
−
x
i
^
∣
∣
2
=
∑
k
=
q
+
1
p
u
k
T
S
u
k
s
.
t
.
u
k
T
u
k
=
1
J = \frac{1}{N} \sum_{i=1}^N ||x_i - \hat{x_i}||^2=\sum_{k=q+1}^p u_k^TSu_k \quad s.t. \quad u_k^Tu_k=1
J=N1∑i=1N∣∣xi−xi^∣∣2=∑k=q+1pukTSuks.t.ukTuk=1
优化问题:
u
k
=
a
r
g
min
∑
k
=
q
+
1
p
u
k
T
S
u
k
s
.
t
.
u
k
T
u
k
=
1
u_k = arg \min \sum_{k=q+1}^p u_k^TSu_k \quad s.t. \quad u_k^Tu_k=1
uk=argmin∑k=q+1pukTSuks.t.ukTuk=1
最大投影方差与最小重构代价的区别
原先由
p
p
p维取前
q
q
q维
最大投影方差目标前
q
q
q最大
最小重构代价目标后
p
−
q
p-q
p−q维方差最小