[Keras]学习笔记(1):optimizers

这篇Keras学习笔记详细介绍了随机梯度下降(SGD),RMSprop,Adagrad,Adadelta,Adam和Adamax等优化器,包括它们的工作原理和参数设置,特别指出这些优化器在深度学习模型训练中的应用和效果。
摘要由CSDN通过智能技术生成

Keras学习笔记(1):optimizers

keras.optimizers.Optimizer()

SGD(Stochastic gradient descent)

随机梯度下降

Stochastic gradient descent, with support for momentum, decay, and Nesterov momentum.

keras.optimizers.SGD(lr=0.01, momentum=0.0, decay=0.0, nesterov=False)
参数
  • lr : float >= 0. Learning rate.学习速率
  • momentum: float >= 0. Parameter updates momentum. 冲量单元
  • decay: float >= 0. Learning rate decay over each update.每次训练学习速率的衰减率
  • nesterov: boolean. Whether to apply Nesterov momentum.Nesterov 冲量单元

RMSprop

RMSProp optimizer.

It is recommended to leave the parameters of this optimizer at their default values.

This optimizer is usually a good choice for recurrent neural networks.

keras.optimizers.Adamax(lr=0.002, beta_1=0.9, beta_2=0.999, epsilon=1e-08)
参数
  • lr : float >= 0. Learning rate.
  • beta_1/beta_2: floats, 0 < beta < 1. Generally close to 1.
  • decay: float >= 0. Learning rate decay over each update.
  • nesterov: boolean. Whether to apply Nesterov momentum.

Adagrad

Adagrad optimizer.

It is recommended to leave the parameters of this optimizer at their default values.

keras.optimizers.Adagrad(lr=0.01, epsilon=1e-06)
参数
  • lr: float >= 0. Learning rate.
  • epsilon: float >= 0.

Adadelta

Adadelta optimizer.

It is recommended to leave the parameters of this optimizer at their default values.

This optimizer is usually a good choice for recurrent neural networks.

keras.optimizers.Adadelta(lr=1.0, rho=0.95, epsilon=1e-06)
参数
  • lr: float >= 0. Learning rate. It is recommended to leave it at the default value.
  • rho: float >= 0.
  • epsilon: float >= 0. Fuzz factor.

Adam

Adam optimizer.

Default parameters follow those provided in the original paper.

keras.optimizers.Adam(lr=0.001, beta_1=0.9, beta_2=0.999, epsilon=1e-08)
参数
  • lr: float >= 0. Learning rate.
  • beta_1/beta_2: floats, 0 < beta < 1. Generally close to 1.
  • epsilon: float >= 0. Fuzz factor.

Adamax

Adam optimizer.

Adamax optimizer from Adam paper’s Section 7. It is a variant of Adam based on the infinity norm.

Default parameters follow those provided in the paper.

keras.optimizers.Adam(lr=0.001, beta_1=0.9, beta_2=0.999, epsilon=1e-08)
参数
  • lr: float >= 0. Learning rate.
  • beta_1/beta_2: floats, 0 < beta < 1. Generally close to 1.
  • epsilon: float >= 0. Fuzz factor.

  • 1
    点赞
  • 2
    收藏
    觉得还不错? 一键收藏
  • 0
    评论
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值