Stanford ML - Lecture 2 - Linear regression with multiple variable

Multiple features


for convenience of notation, define , the new hypothesis is

Gradient descent for multiple variables
new algorithm is


Gradient descent in practive I: Feature Scaling
make sure features are on a similar scale
  • feature scaling

  • mean normalization

Gradient descent in practice II: Learning rate
  • make sure gradient descent is working correctly
  • gradient descent not working - use smaller \alpha
  • to choose , try ..., 0.001, 0.003, 0.01, 0.03, ..., 0.1, ..., 1, ...
Features and polynomial regression

choice of features



Normal equations

method for solve analytically

if there are m examples, each example have n input features, so is (n+1)-by-1 matrix, X is m-by-(n+1) matrix, y is m-dimensional vector

                                   ?






Advantages and disadvantages of gradient descent and normal equation

 training examples,  features

Gradient descent

  • need to choose 
  • need many iterations
  • works well even when  is large
Normal equation
  • no need to choose 
  • don't need to iterate
  • need to compute 
  • slow if  is very large
  • if  is larger than 10000 or 100000, need consider to choose gradient descent
  • only suitable for linear regression
Normal equation and non-invertibility (optional)
  • redundant features (linearly dependent)
  • too many features (e.g. )
    • delete some features, or use regularization
  • 0
    点赞
  • 0
    收藏
    觉得还不错? 一键收藏
  • 0
    评论
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值