2020-9-7 吴恩达-改善深层NN-w3 超参数调试、Batch正则化和程序框架(课后作业)

参考链接

1、If searching among a large number of hyperparameters, you should try values in a grid rather than random values, so that you can carry out the search more systematically and not rely on chance. True or False?
如果在大量的超参数中搜索,那么应该尝试在网格中搜索而不是使用随机值,以便更系统的搜索,而不是依靠运气。

答案:错误

And to take an extreme example, let’s say that hyperparameter two was that value epsilon that you have in the denominator of the Adam algorithm. So your choice of alpha matters a lot and your choice of epsilon hardly matters.
应当尝试随机值,不要使用网格搜索,因为你不知道哪些超参数比其他的更重要。
举一个很极端的例子,就比如在Adam算法中参数2-防止除零操作的ε的值,一般为1的负8次方,但是和学习率α相比,ε就显得不那么重要了。

===========================================================
2、Every hyperparameter, if set poorly, can have a huge negative impact on training, and so all hyperparameters are about equally important to tune well. True or False?
每个超参数如果设置得不好,都会对训练产生巨大的负面影响。因此调整好所有的超参数都很重要,请问这是正确的吗?

答案:错误

We’ve seen in lecture that some hyperparameters, such as the learning rate, are more critical than others.
我们在课程中讲到过,有些参数,比如学习率就比其他的超参数更加重要。

===========================================================
3、During hyperparameter search, whether you try to babysit one model (“Panda” strategy) or train a lot of models in parallel (“Caviar”) is largely determined by:
在超参数搜索过程中,你尝试只照顾一个模型(使用熊猫策略)还是一起训练大量的模型(鱼子酱策略)在很大程度上取决于:

  • Whether you use batch or mini-batch optimization
  • The presence of local minima (and saddle points) in your neural network
  • The amount of computational power you can access 你拥有的计算能力。正确。
  • The number of hyperparameters you have to tune

课程链接

===========================================================
4、If you think β (hyperparameter for momentum) is between on 0.9 and 0.99, which of the following is the recommended way to sample a value for beta?
如果你认为β(动量超参数)介于0.9和0.99之间,那么推荐采用以下哪一种方法来对β值进行取样?

  • r = np.random.rand()
    beta = r*0.09+0.9
  • r = np.random.rand()
    beta = 1 - 10 ** (-r - 1) 正确
  • r = np.random.rand()
    beta = 1 - 10 ** (r + 1)
  • r = np.random.rand()
    beta = r*0.9+0.09

参考链接

=============================================

  • 0
    点赞
  • 0
    收藏
    觉得还不错? 一键收藏
  • 0
    评论

“相关推荐”对你有帮助么?

  • 非常没帮助
  • 没帮助
  • 一般
  • 有帮助
  • 非常有帮助
提交
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值