Stanford ML - Lecture 6 - Advice for applying machine learning

1. Deciding what to try next

  • Debugging a learning algorithm
    • Suppose you have implemented regularized linear regression to predict housing prices, when you test your hypothesis on a new set of houses, you find that it makes unacceptably large errors in its predictions. What should you try next?
      1. Get more training examples
      2. Try smaller sets of features
      3. Try getting additional features
      4. Try adding polynomial features
      5. Try decreasing 
      6. Try increasing 

2. Evaluating a hypothesis

  • separate data sets into training set (70%) andtest set (30%)
  • Training/Testing procedure for logistic regression
    • learn parameter  from training data
    • compute test set error
    • misclassification error (0/1 misclassification error)

3. Model selection and training/validation/test sets

  • overfitting example
    • the training error is likely to be lower than the actual generalization error
  • model selection
    • select the model that has the lowest test error
  • training set - 60%
  • cross validation set (cv) - 20%
  • test set - 20%
  • training error

  • cross validation error

  • test error
4. Diagnosing bias vs. variance

  • bias (underfit)


  • variance (overfit)

5. Regularization and bias/variance

  • choosing the regularization parameter 


    • ........

6. Learning curves

  • If a learning algorithm is suffering from high bias, getting more training data will not (by itself) help much.
  • If a learning algorithm is suffering from high variance, getting more training data is likely to help.
7. Deciding what to try next (revisited)

  • "small" neural network (fewer parameters, more prone to underfitting)
    • computationally cheapter
  • "large" neural network (more parameters, more prone to overfitting)
    • computationally more expensive
    • use regularization to address overfitting

the definition:

Variance: measures the extent to which the solutions for individual data sets vary around their average, hence this measures the extent to which the function f(x) is sensitive to theparticular choice of data set.

Bias: represents the extent to which the average prediction over all data sets differs from the desired regression function.

variance:估计本身的方差。

bias:估计的期望和样本数据样本希望得到的回归函数之间的差别。

From : http://blog.csdn.net/abcjennifer/article/details/7797502

  • 0
    点赞
  • 0
    收藏
    觉得还不错? 一键收藏
  • 0
    评论
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值