参考:tensorflow实战 黄文坚
# *_*coding:utf-8 *_*
from datetime import datetime
import math
import time
import tensorflow as tf
from tensorflow.examples.tutorials.mnist import input_data
mnist = input_data.read_data_sets('MNIST_data', one_hot=True)
#卷积层函数
'''
kh、kw卷积核的高和宽,n_out卷积核数量即输出通道数,dh、dw步长的长和宽, p=参数列表
'''
def conv_op(input_op, name, kh, kw, n_out, dh, dw, p):
#get_shape()[-1].value 获取输入input_op的通道数,比如224x224x3中最后那个3
n_in = input_op.get_shape()[-1].value
with tf.name_scope(name) as scope:
'''tf.contrib.layers.xavier_initializer_conv2d根据某一层网络的输出输入节点数量自动调节最合适的分布'''
kernel = tf.get_variable(scope + 'w', shape=[kh, kw, n_in, n_out], dtype=tf.float32,
initializer=tf.contrib.layers.xavier_initializer_conv2d())
conv = tf.nn.conv2d(input_op, kernel, (1, dh, dw, 1), padding='SAME')
bias_init_val = tf.constant(0.0, shape=[n_out], dtype=tf.float32)
biases = tf.Variable(bias_init_val, trainable=True, name='b')
z = tf.nn.bias_add(conv, biases)
activation = tf.nn.relu(z, name=scope)
p += [kernel, biases]
return activation
#全连接层函数
def fc_op(input_op, name, n_out, p):
n_in = input_op.get_shape()[-1].value
with tf.name_scope(name) as scope:
kernal = tf.get_variable(scope + 'w', shape=[n_in, n_out], dtype=tf.float32,
initializer=tf.contrib.layers.xavier_initializer())
#这里biases不在初始化为0,而赋予一个较小的值0.1以避免dead neuron(神经死亡)
biases = tf.Variable(tf.constant(0.1, shape=[n_out], dtype=tf.float32), name='b')
activation = tf.nn.relu_layer(input_op, kernal, biases, name=scope)
p += [kernal, biases]
return activation
#创建最大池化层函数
def maxpool_op(input_op, name, kh, kw, dh, dw):
return tf.nn.max_pool(input_op, ksize=[1, kh, kw, 1], strides=[1, dh, dw, 1], padding='SAME', name=name)
#前向传播函数
def inference_op(input_op, keep_prob):
p = []