Notes for Deep Learning Lessons of Pro. Hung-yi Lee (2)

Today, Knowledge concerning about the optimization of deep learning is written here. What is the meaning of optimaztion? The following ppt shows us the answer.

在这里插入图片描述

1. SGD with Momentum (SGDM)

Just as the name shows us, SGDM is invented by combining SGD with Momentum. For SGDM, the process of updating parameters is shown in the following ppt. What we should pay attention to, or in the other words what makes us better understand the meanings of SGDM, is v i v^i vi is actually the weighted sum of all the previous gradient and the closer gradient has more influence on current momentum.
在这里插入图片描述
What is the adavantage of adding momentum in SGD program? For SGD, it is easy to lead us to local minima point rather than the global minima point. However, adding momentum takes the history information into account, which means, if we explains it in a more vivid way, SGDM offers us the ability to think whether we are standing at the local minima point.
在这里插入图片描述

2. Adagrad

It has been introduced in last blog. So, I do not want to explain it again. (I am a little bit lazy, haha).
在这里插入图片描述

3. RMSProp

RMSProp makes a little change on the formula of Adagrad. In Adagrad, v t v_t vt is the sum of square of the past gradient. However, in RMSProp, v t = α v t − 1 + ( 1 − α ) ( g t − 1 ) 2 v_t = \alpha v_{t-1}+(1-\alpha)(g_{t-1})^2 vt=αvt1+(1α)(gt1)2. We can changes the value of α \alpha α to make v t − 1 v_{t-1} vt1 have more or less influence on the current gradient. In common situation, we always set α \alpha α as a large number with the afraid of a too large g t − 1 g_{t-1} gt1 making η v t \frac{\eta}{\sqrt{v_t}} vt η too close to zero.

在这里插入图片描述

4. Adam

If we ignore some little differences, Adam can be seen as the combination of SGDM and RMSProp. The little change is that we change the form of m t m_t mt, which can be called as de-biasing. The reason of this change is that the value of m t m_t mt is too close to zero at the beginning of updating.

在这里插入图片描述

  • 1
    点赞
  • 0
    收藏
    觉得还不错? 一键收藏
  • 0
    评论
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值