参考 keras.optimizers - 云+社区 - 腾讯云
Built-in optimizer classes.
Modules
- schedules module: Public API for tf.keras.optimizers.schedules namespace.
Classes
- class Adadelta: Optimizer that implements the Adadelta algorithm.
- class Adagrad: Optimizer that implements the Adagrad algorithm.
- class Adam: Optimizer that implements the Adam algorithm.
- class Adamax: Optimizer that implements the Adamax algorithm.
- class Ftrl: Optimizer that implements the FTRL algorithm.
- class Nadam: Optimizer that implements the NAdam algorithm.
- class Optimizer: Updated base class for optimizers.
- class RMSprop: Optimizer that implements the RMSprop algorithm.
- class SGD: Stochastic gradient descent and momentum optimizer.
Functions
- deserialize(...): Inverse of the
serialize
function. - get(...): Retrieves a Keras Optimizer instance.
- serialize(...)