RNN and its mathematical deduction

RNN(Recurrent Neural Networks)

1. Introduction

Today, We are going to talk about another Neural Network, we call it RNN(Recurrent Neural Networks).This type of network has many significant meanings. Recurrent Neural Network is very effective for data with sequence rows, and it can mine the temporal information and semantic information in the data.

2. Topology

在这里插入图片描述

RNN is mostly similar to BP networks, however, it adds a recurrent layer to its hidden layer. When we unfold it, we could find its special structure. This structure determines that it can memorize the former information. According to its characteristic, we can implement statement recognition.

Mathematic deduction

First, we are supposed to introduce a few quantities.

在这里插入图片描述
Above is a neuronal node, we call it S1. In the bottom of this node, there is h1, and its mathematic expression is h1 = UX1 + WS0. And in the top of this node, we may find an f(x), we call is activation function, this function may be sigmoid, RELU, TANH and so on .

Then we will do our deduction(This deduction we call it BPTT algorithm ):

  1. find the gradient of each parameter:
    在这里插入图片描述
    在这里插入图片描述
    在这里插入图片描述

  2. find the update method of w:
    在这里插入图片描述

And, the third equation, we could express as:
在这里插入图片描述

(Hints: The plus sign on the exponent in this figure means that the derivation is only related to S3)

In the same way: the third equation of the above expression could express as :
在这里插入图片描述

In the same way: 在这里插入图片描述

Then let us integrate these three equations:

在这里插入图片描述

In fact, we are able to introduce a more simpler expression to reface this equation:
在这里插入图片描述

In the same way, other two parameters’ update method we could conclude as:
在这里插入图片描述
在这里插入图片描述

Above are the specific deduction of RNN.

Conclusion:

RNN is a useful algorithm that could solve semantic information problems. But this is not enough. As we can see in its topology, the number of nodes of recurrent layer above are just three. When we need to add more nodes to its layer, it may be not as effective as we are supposed to because it could lost lots of information. Hence, in order to handle these problems, we are supposed to introduce another algorithm called LTMS algorithms which we will talk in my next blog.

If you guys want to know more about machine learning, PLZ follow me.
I will update more articles these days.

  • 3
    点赞
  • 0
    收藏
    觉得还不错? 一键收藏
  • 0
    评论
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值