Keras手册记录之optimizer, 待更新各个算法详解!

2 篇文章 0 订阅

optimizer:https://keras.io/optimizers/

下面recommended to leave the parameters of this optimizer at their default values的优化算法,名字旁边打星号

SGD

 keras.optimizers.SGD(lr=0.01, momentum=0.0, decay=0.0, nesterov=False)

RMSprop 适用于RNN*

keras.optimizers.RMSprop(lr=0.001, rho=0.9, epsilon=None, decay=0.0)

链接:http://www.cs.toronto.edu/~tijmen/csc321/slides/lecture_slides_lec6.pdf

Adagrad*

keras.optimizers.Adagrad(lr=0.01, epsilon=None, decay=0.0)

Adagrad is an optimizer with parameter-specific learning rates, which are adapted relative to how frequently a parameter gets updated during training. The more updates a parameter receives, the smaller the learning rate.

链接:http://www.jmlr.org/papers/volume12/duchi11a/duchi11a.pdf

Adadelta,a more robust extension of Adagrad*

keras.optimizers.Adadelta(lr=1.0, rho=0.95, epsilon=None, decay=0.0)

链接:https://arxiv.org/abs/1212.5701

Adam

keras.optimizers.Adam(lr=0.001, beta_1=0.9, beta_2=0.999, epsilon=None, decay=0.0, amsgrad=False)

链接:

https://arxiv.org/abs/1412.6980v8

https://openreview.net/forum?id=ryQu7f-RZ

Adamax,a variant of Adam based on the infinity norm

keras.optimizers.Adamax(lr=0.002, beta_1=0.9, beta_2=0.999, epsilon=None, decay=0.0)

链接同adam

Nadam*,Much like Adam is essentially RMSprop with momentum, Nadam is Adam RMSprop with Nesterov momentum.

keras.optimizers.Nadam(lr=0.002, beta_1=0.9, beta_2=0.999, epsilon=None, schedule_decay=0.004)

链接:

http://cs229.stanford.edu/proj2015/054_report.pdf

http://www.cs.toronto.edu/~fritz/absps/momentum.pdf

 

  • 2
    点赞
  • 9
    收藏
    觉得还不错? 一键收藏
  • 0
    评论
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值