RNN & Generative Model

LSTM

  • With gated RNN,the network learns to which info is remembered and which should be forgot over a long duration (through forgot gate).
  • distinguish between the cell state and hidden state, the former aims to maintain a long term dependency, the latter is just the input of (forget, input and gate gate) and output of output gate.
  • The introduction of cell state in LSTM is the primary reason why the vanishing or exploding gradient is mitigated. Pls see Tutorial here.

Others

  • Eg of image captioning:
    • combination of CNN and RNN
    • CNN takes the input of an image and output a feature vector
    • then this feature vector is input into RNN as something like a hidden state (but actually not!!), with conversion matrix: Wih W i h
  • Gradient cliping:
    • solving two problems: sharp cliff in parameter space and exploding gradient space.

The basic idea is to recall that the gradient specifies not the optimal step size, but only the optimal direction within an infinitesimal region.
The objective function for highly nonlinear deep neural networks or forrecurrent neural networks often contains sharp nonlinearities in parameter space resulting from the multiplication of several parameters.
Thus limit the gradient size by a predefined threshold.

  • Exploding and vanishing gradient:
    • It is sufficient for λ1<1γ λ 1 < 1 γ for the vanishing gradient occurs.
    • The necessary condition for exploding gradient is the largest singular value λ1>1γ λ 1 > 1 γ

Tutorial

Generative Model:

  • Training example of GAN:
    • we sample a mini-batch of m noise example { z(1),,z<
  • 0
    点赞
  • 0
    收藏
    觉得还不错? 一键收藏
  • 0
    评论

“相关推荐”对你有帮助么?

  • 非常没帮助
  • 没帮助
  • 一般
  • 有帮助
  • 非常有帮助
提交
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值