keras example5——mnist_hierarchical_rnn

# -*- coding:utf-8 -*-

import keras
from keras.datasets import mnist
from keras.models import Model
from keras.layers import Input, Dense, TimeDistributed, LSTM

(x_train, y_train), (x_test, y_test) = mnist.load_data()

x_train = x_train.reshape(x_train.shape[0], 28, 28, 1).astype('float32')/255
x_test = x_test.reshape(x_test.shape[0], 28, 28, 1).astype('float32')/255

y_train = keras.utils.to_categorical(y_train, 10)
y_test = keras.utils.to_categorical(y_test, 10)

row, col, pixel = x_train.shape[1:]  # 28,28,1

x = Input(shape=(row, col, pixel))
encoded_rows = TimeDistributed(LSTM(128))(x)  # Encodes a row of pixels using TimeDistributed Wrapper.
encoded_cols = LSTM(128)(encoded_rows)  # Encodes columns of encoded rows.
prediction = Dense(10, activation='softmax')(encoded_cols)
model = Model(x, prediction)
model.summary(line_length=100)

model.compile(loss=keras.losses.categorical_crossentropy,
              optimizer=keras.optimizers.rmsprop(),
              metrics=['accuracy'])

model.fit(x_train, y_train, batch_size=32,
          epochs=10, verbose=1, validation_data=(x_test, y_test))

scores = model.evaluate(x_test, y_test, verbose=1)
print(scores)

# loss: 0.0287 - acc: 0.9911 - val_loss: 0.0395 - val_acc: 0.9887

"""
____________________________________________________________________________________________________
Layer (type)                                 Output Shape                            Param #
====================================================================================================
input_1 (InputLayer)                         (None, 28, 28, 1)                       0
____________________________________________________________________________________________________
time_distributed_1 (TimeDistributed)         (None, 28, 128)                         66560
____________________________________________________________________________________________________
lstm_2 (LSTM)                                (None, 128)                             131584
____________________________________________________________________________________________________
dense_1 (Dense)                              (None, 10)                              1290
====================================================================================================
Total params: 199,434
Trainable params: 199,434
Non-trainable params: 0
____________________________________________________________________________________________________
Train on 60000 samples, validate on 10000 samples
"""

RNN主要用在自然语言处理上,这是个图像分类的例子。

  • 0
    点赞
  • 0
    收藏
    觉得还不错? 一键收藏
  • 0
    评论
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值