使用Tensorflow 2.2 中的keras实现双向循环神经网络。训练速度非常慢。第一层的单元是64,第二层是32。通过model.summary()可以发现,实际上每一层都是乘以2的。
import tensorflow as tf
import numpy as np
from tensorflow import keras
import os
os.environ['TF_CPP_MIN_LOG_LEVEL'] = '1'
#读取本地mnist数据
def my_load_data(path='mnist.npz'):
origin_folder = 'https://storage.googleapis.com/tensorflow/tf-keras-datasets/'
path = tf.keras.utils.get_file(
path,
origin=origin_folder + 'mnist.npz',
cache_dir='DataSet/',
cache_subdir=""
)
with np.load(path, allow_pickle=True) as f:
x_train, y_train = f['x_train'], f['y_train']
x_test, y_test = f['x_test'], f['y_test']
return (x_train, y_train), (x_test, y_test)
(x_train, y_train), (x_test, y_test) = my_load_data(path='mnist.npz')
x_train=x_train/255.
x_test=x_test/255.
#转换为独热模式
y_train = keras.utils.to_categorical(y_train, num_classes=10)
y_test = keras.utils.to_categorical(y_test, num_classes=10)
# 数据长度 一行有28个像素
input_size = 28
# 序列的长度
time_steps = 28
# 隐藏层block的个数
cell_size = 64
b_size = 32
model = keras.Sequential()
# 循环神经网络
model.add(keras.layers.Bidirectional(
keras.layers.SimpleRNN(cell_size, return_sequences=True),
input_shape = (time_steps, input_size)
))
model.add(keras.layers.Bidirectional(keras.layers.SimpleRNN(b_size)))
# 输出层
model.add(keras.layers.Dense(10, activation='softmax'))
# 定义优化器
adam = keras.optimizers.Adam(lr=1e-4)
# 定义优化器、loss function, 训练过程中计算准确率
model.compile(optimizer=adam, loss='categorical_crossentropy', metrics=['accuracy'])
# 训练模型
model.fit(x_train, y_train, batch_size=64, epochs=3)
# 评估模型
loss, accuracy = model.evaluate(x_test, y_test)
print('test loss', loss)
print('test accuracy', accuracy)