Deep Learning:Optimization for Training Deep Models(零)

Of all of the many optimization problems involved in deep learning, the most difficult is neural network training.
It is quite common to invest days to months of time on hundreds of machines in order to solve even a single instance of the neural network training problem.
Because this problem is so important and so expensive, a specialized set of optimization techniques have been developed for solving it. This chapter presents these optimization techniques for neural network training.
This chapter focuses on one particular case of optimization: finding the parameters θ of a neural network that significantly reduce a cost function J(θ), which typically includes a performance measure evaluated on the entire training set as well as additional regularization terms.

  • We begin with a description of how optimization used as a training algorithm for a machine learning task differs from pure optimization.
  • Next, we present several of the concrete challenges that make optimization of neural networks difficult.
  • We then define several practical algorithms, including both optimization algorithms themselves and strategies for initializing the parameters. More advanced algorithms adapt their learning rates during training or leverage information contained in the second derivatives of the cost function.
  • Finally, we conclude with a review of several optimization strategies that are formed by combining simple optimization algorithms into higher-level procedures.
  • 0
    点赞
  • 0
    收藏
    觉得还不错? 一键收藏
  • 0
    评论
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值