梯度下降法

梯度下降法

随机梯度下降法

批梯度下降法

g k = f k ′ ( x ) g_k = f_k'(x) gk=fk(x)
x k + 1 = x k − l r ∗ g k x_{k+1} = x_k - lr * g_k xk+1=xklrgk

小批梯度下降法

优化算法

动量

g k = f k ′ ( x ) g_k = f_k'(x) gk=fk(x)
v k = α v k − 1 − l r ∗ g k v_k = \alpha v_{k-1} -lr*g_k vk=αvk1lrgk
x k + 1 = x k + v k = x k + α v k − l r ∗ g k − 1 x_{k+1} = x_k +v_k = x_k+\alpha v_k - lr*g_{k-1} xk+1=xk+vk=xk+αvklrgk1

Nesterow 动量

g k = f k ′ ( x + α v k ) g_k = f_k'(x+\alpha v_k) gk=fk(x+αvk)
v k = α v k − 1 − l r ∗ g k v_k = \alpha v_{k-1} -lr*g_k vk=αvk1lrgk
x k + 1 = x k + v k = x k + α v k − l r ∗ g k − 1 x_{k+1} = x_k +v_k = x_k+\alpha v_k - lr*g_{k-1} xk+1=xk+vk=xk+αvklrgk1

AdaGrad

g k = f k ′ ( x ) g_k = f_k'(x) gk=fk(x)
r k = r k − 1 + g ∗ g r_k = r_{k-1}+g*g rk=rk1+gg
x k + 1 = x k − l r δ + r k ∗ g k x_{k+1} = x_k - \frac{lr}{\delta+\sqrt r_k} * g_k xk+1=xkδ+r klrgk

RMSProp

g k = f k ′ ( x ) g_k = f_k'(x) gk=fk(x)
r k = ρ r k − 1 + ( 1 − ρ ) g ∗ g r_k = \rho r_{k-1}+(1-\rho)g*g rk=ρrk1+(1ρ)gg
x k + 1 = x k − l r δ + r k ∗ g k x_{k+1} = x_k - \frac{lr}{\delta+\sqrt r_k} * g_k xk+1=xkδ+r klrgk

Adam

g k = f k ′ ( x ) g_k = f_k'(x) gk=fk(x)
s k = ρ 1 s k − 1 + ( 1 − ρ 1 ) g s_k = \rho_1 s_{k-1}+(1-\rho_1)g sk=ρ1sk1+(1ρ1)g
r k = ρ 2 r k − 1 + ( 1 − ρ 2 ) g ∗ g r_k = \rho_2 r_{k-1}+(1-\rho_2)g*g rk=ρ2rk1+(1ρ2)gg
s ^ = s 1 − ρ 1 k \hat s = \frac {s}{1-\rho_1^k} s^=1ρ1ks
r ^ = r 1 − ρ 2 k \hat r = \frac {r}{1-\rho_2^k} r^=1ρ2kr
x k + 1 = x k − l r s ^ k δ + r ^ k x_{k+1} = x_k - \frac{lr \hat s_k}{\delta+\sqrt {\hat r_k}} xk+1=xkδ+r^k lrs^k

  • 0
    点赞
  • 1
    收藏
    觉得还不错? 一键收藏
  • 0
    评论

“相关推荐”对你有帮助么?

  • 非常没帮助
  • 没帮助
  • 一般
  • 有帮助
  • 非常有帮助
提交
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值