机器学习笔记 ---- Principal Component Analysis

1. Task of PCA

Find a direction and project all points to that line, thus minimizing the projection error.
Projection error: Sum of distances between points and line

2. Data Preprocessing

Feature Scaling + Mean Normalization

3. PCA Algorithm



Using the first k vectors in U U and denote it as Ur , the result is Z=UTrX Z = U r T X

4. Reconstruction from PCA


Xapprox=UrZ X a p p r o x = U r Z

5. How to Choose the Reduced Dimension



Using S=diag(s1...sn) S = d i a g ( s 1 . . . s n ) , Check whether

1ki=1sini=1si<=0.01 1 − ∑ i = 1 k s i ∑ i = 1 n s i <= 0.01

—– an O(n) O ( n ) Algorithm

6. Speed Up Supervised Learning by PCA

Train the model using data compressed by PCA
Note: Running PCA which only depends on TRAINING SET when training!
While this mapping can be applied to other sets.

Only use PCA when the original data perform badly on your system!
  • 0
    点赞
  • 0
    收藏
    觉得还不错? 一键收藏
  • 0
    评论
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值