Introduction to Recurrent Neural Networks

What is RNN

The networks are recurrent because they performance same computations for all the elements of a sequence of input, and the output of each element dependents, in addition to current input, from all the previous commutations.

Why RNN

  • Sequential type information of the inputs
    Video Analysis
    Speech Recognition
    Machine Translation
  • RNN have proved to have excellent performance in such problems

RNN Procedure

这里写图片描述

Sigmoid Gradient

这里写图片描述

The Vanish Gradient Problem

Consider the recurrent networks:

ht=σ(Uxt+Vht1) h t = σ ( U x t + V h t − 1 )

then,
h3=σ(Ux3+V(σ(Ux2+V(σ(Ux1))))) h 3 = σ ( U x 3 + V ( σ ( U x 2 + V ( σ ( U x 1 ) ) ) ) )

E3U=E3out3out3h3h3h2h2h1h1U ∂ E 3 ∂ U = ∂ E 3 ∂ o u t 3 ∂ o u t 3 ∂ h 3 ∂ h 3 ∂ h 2 ∂ h 2 ∂ h 1 ∂ h 1 ∂ U

LSTM Cell

这里写图片描述

  • Input Gate

    g=tanh(bg+xtUg+ht1Vg) g = t a n h ( b g + x t U g + h t − 1 V g )

    i=σ(bi+xtUi+ht1Vi) i = σ ( b i + x t U i + h t − 1 V i )

    outi=gi o u t i = g ∘ i

  • forget gate

    f=σ(bf+xtUf+ht1Vf) f = σ ( b f + x t U f + h t − 1 V f )

    st=st1f+gi s t = s t − 1 ∘ f + g ∘ i

  • output gate

    o=σ(bo+xtUo+ht1Vo) o = σ ( b o + x t U o + h t − 1 V o )

    ht=tanh(st)o h t = t a n h ( s t ) ∘ o

Reducing The Problem

stst1=f ∂ s t ∂ s t − 1 = f

Reference

  • 1
    点赞
  • 0
    收藏
    觉得还不错? 一键收藏
  • 0
    评论
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值