本文参考了https://blog.csdn.net/sysstc/article/details/74910918 的代码,小弟我在照着代码抄的时候,发现CNN和DNN部分代码存在问题,感谢@weixin_39739789 的提示,我把自己修改后的代码发上来。这些代码在Ubuntu16.04下运行无误哈。
CNN
from tensorflow.examples.tutorials.mnist import input_data
mnist = input_data.read_data_sets("mnist_data", one_hot=True)
import tensorflow as tf
learning_rate = 0.001
training_iters = 200000
batch_size = 128
display_step = 20
n_input = 784
n_classes = 10
drop_out = 0.8
x = tf.placeholder(tf.float32, [None, n_input])
y = tf.placeholder(tf.float32, [None, n_classes])
keep_prob = tf.placeholder(tf.float32)
def conv2d(name, l_input, w, b):
return tf.nn.relu(tf.nn.bias_add(tf.nn.conv2d(l_input, w, strides=[1, 1, 1, 1],padding='SAME'), b), name = name)
def max_pool(name,l_input, k):
return tf.nn.max_pool(l_input, ksize=[1, k, k, 1], strides=[1, k, k, 1], padding='SAME', name=name)
def mynorm(name, l_input, lsize=4):
return tf.nn.lrn(l_input, lsize, bias=1.0, alpha=0.001/90, beta=0.75, name=name)
def customnet(_X, _weights, _biases, _dropout):
_X = tf.reshape(_X, shape=[-1, 28, 28, 1])
conv1 = conv2d('conv1', _X, _weights['wc1'], _biases['bc1'])
pool1 = max_pool('conv1', conv1, k=2)
norm1 = mynorm('conv1', pool1, lsize=4)
norm1 = tf.nn.dropout(norm1, _dropout)
conv2 = conv2d('conv2', norm1, _weights['wc2'], _biases['bc2'])
pool2 = max_pool('conv2', conv2, k=2)
norm2 = mynorm('conv2', pool2, lsize=4)
norm2 = tf.nn.dropout(norm2, _dropout)
conv3 = conv2d('conv3', norm2, _weights['wc3'], _biases['bc3'])
pool3 = max_pool('conv3', conv3, k=2)
norm3 = mynorm('conv3', pool3, lsize=4)
norm3 = tf.nn.dropout(norm3, _dropout)
conv4 = conv2d('conv4', norm3, _weights['wc4'], _biases['bc4'])
pool4 = max_pool('conv4', conv4, k=2)
norm4 = mynorm('conv4', pool4, lsize=4)
norm4 = tf.nn.dropout(norm4, _dropout)
dense1 = tf.reshape(norm4, [-1, _weights['wd1'].get_shape().as_list()[0]])
dense1 = tf.nn.relu(tf.matmul(dense1, _weights['wd1']) + _biases['bd1'], name='fc1')
dense2 = tf.nn.relu(tf.matmul(dense1, _weights['wd2']) + _biases['bd2'], name='fc2')
out = tf.matmul(dense2, _weights['out']) + _biases['out']
return out
weights = {
'wc1': tf.Variable(tf.random_normal([5, 5, 1, 64])),
'wc2': tf.Variable(tf.random_normal([5, 5, 64, 128])),
'wc3': tf.Variable(tf.random_normal([3, 3, 128, 256])),
'wc4': tf.Variable(tf.random_normal([2, 2, 256, 512])),
'wd1': tf.Variable(tf.random_normal([2*2*512, 1024])),
'wd2': tf.Variable(tf.random_normal([1024, 1024])),
'out': tf.Variable(tf.random_normal([1024, 10]))
}
biases = {
'bc1': tf.Variable(tf.random_normal([64])),
'bc2': tf.Variable(tf.random_normal([128])),
'bc3': tf.Variable(tf.random_normal([256])),
'bc4': tf.Variable(tf.random_normal([512])),
'bd1': tf.Variable(tf.random_normal([1024])),
'bd2': tf.Variable(tf.random_normal([1024])),
'out': tf.Variable(tf.random_normal([n_classes])),
}
pred = customnet(x, weights, biases, keep_prob)
cost = tf.reduce_mean(tf.nn.softmax_cross_entropy_with_logits(labels=y, logits=pred))
optimizer = tf.train.AdamOptimizer(learning_rate=learning_rate).minimize(cost)
correct_pred = tf.equal(tf.argmax(pred, 1), tf.argmax(y, 1))
accuracy = tf.reduce_mean(tf.cast(correct_pred, tf.float32))
init = tf.initialize_all_variables()
with tf.Session() as sess:
sess.run(init)
step = 1
while step * batch_size < training_iters:
batch_xs, batch_ys = mnist.train.next_batch(batch_size)
sess.run(optimizer, feed_dict={x: batch_xs, y: batch_ys, keep_prob: drop_out})
if step % display_step == 0:
acc = sess.run(accuracy, feed_dict={x: batch_xs, y: batch_ys, keep_prob: 1.})
loss = sess.run(cost, feed_dict={x: batch_xs, y: batch_ys, keep_prob: 1.})
print("Iter" + str(step * batch_size) + ", Minibatch Loss=" + "{:.6f}".format(loss) + ", Training Accuracy=" + "{:.5f}".format(acc))
step += 1
print("Optimization Finished!!!")
print("Testing Accuracy:", sess.run(accuracy, feed_dict={x: mnist.test.images[:256], y: mnist.test.labels[:256], keep_prob: 1.}))
DNN
import tensorflow as tf
from tensorflow.examples.tutorials.mnist import input_data
mnist = input_data.read_data_sets("mnist_data", one_hot=True)
learning_rate = 0.001
#training_iters = 100000
training_iters = 400000
batch_size = 128
display_step = 10
n_input = 784
n_classes = 10
dropout = 0.8
keep_prob = tf.placeholder(tf.float32)
x = tf.placeholder(tf.float32, [None, n_input])
y = tf.placeholder(tf.float32, [None, n_classes])
def conv2d(image, w, b):
return tf.nn.relu(tf.nn.bias_add(tf.nn.conv2d(image, w, strides=[1, 1, 1, 1], padding="SAME"), b))
def max_pool(image, k):
return tf.nn.max_pool(image, ksize=[1, k, k, 1], strides=[1, k, k, 1], padding="SAME",)
def dnn(_X,_weights, _biases, _dropout):
_X = tf.nn.dropout(_X, _dropout)
d1 = tf.nn.relu(tf.nn.bias_add(tf.matmul(_X, _weights['wd1']), _biases['bd1']), name='d1')
d2x = tf.nn.dropout(d1, _dropout)
d2 = tf.nn.relu(tf.nn.bias_add(tf.matmul(d2x, _weights['wd2']), _biases['bd2']), name='d2')
dout = tf.nn.dropout(d2, _dropout)
out = tf.matmul(dout, _weights['out']) + _biases['out']
return out
weights = {
'wd1': tf.Variable(tf.random_normal([784, 600], stddev=0.01)),
'wd2': tf.Variable(tf.random_normal([600, 480], stddev=0.01)),
'out': tf.Variable(tf.random_normal([480, 10]))
}
biases = {
'bd1': tf.Variable(tf.random_normal([600])),
'bd2': tf.Variable(tf.random_normal([480])),
'out': tf.Variable(tf.random_normal([10]))
}
pred = dnn(x, weights, biases, keep_prob)
cost = tf.reduce_mean(tf.nn.softmax_cross_entropy_with_logits(labels=y, logits=pred))
optimizer = tf.train.AdamOptimizer(learning_rate=learning_rate).minimize(cost)
correct_pred = tf.equal(tf.argmax(pred, 1), tf.argmax(y, 1))
accuracy = tf.reduce_mean(tf.cast(correct_pred, tf.float32))
init = tf.initialize_all_variables()
with tf.Session() as sess:
sess.run(init)
step = 1
while step * batch_size < training_iters :
batch_xs, batch_ys = mnist.train.next_batch(batch_size)
sess.run(optimizer, feed_dict={x: batch_xs, y: batch_ys, keep_prob: dropout})
if step % display_step == 0:
acc = sess.run(accuracy, feed_dict={x: batch_xs, y: batch_ys, keep_prob: 1.})
loss = sess.run(cost, feed_dict={x: batch_xs, y: batch_ys, keep_prob: 1.})
print("Iter " + str(step * batch_size) + "Minibatch Loss=" + "{:.6f}".format(loss) +
", Training Accuracy=" + "{:.5f}".format(acc))
step += 1
print("Optimization Finished!!!!!!")
print("Testing Accuracy:",
sess.run(accuracy, feed_dict={x: mnist.test.images[:256], y: mnist.test.labels[:256], keep_prob: 1.}))
ANN
import tensorflow as tf
from tensorflow.examples.tutorials.mnist import input_data
from pyexpat import features
mnist = input_data.read_data_sets("mnist_data", one_hot=True)
learning_rate = 0.001
training_epochs = 150 # origin epochs = 15
batch_size = 100
display_step = 1
n_hidden_1 = 256
n_hidden_2 = 256
n_input = 784
n_classes = 10
x = tf.placeholder("float", [None, 784])
y = tf.placeholder("float", [None, 10])
def multilayer_perception(_X, _weights, _biases):
layer1 = tf.nn.sigmoid(tf.add(tf.matmul(_X, _weights['h1']),_biases['b1']))
layer2 = tf.nn.sigmoid(tf.add(tf.matmul(layer1, _weights['h2']), _biases['b2']))
return tf.matmul(layer2, _weights['out']) + _biases['out']
weights = {
'h1': tf.Variable(tf.random_normal([n_input, n_hidden_1])),
'h2': tf.Variable(tf.random_normal([n_hidden_1, n_hidden_2])),
'out': tf.Variable(tf.random_normal([n_hidden_2, n_classes]))
}
biases = {
'b1': tf.Variable(tf.random_normal([n_hidden_1])),
'b2': tf.Variable(tf.random_normal([n_hidden_2])),
'out': tf.Variable(tf.random_normal([n_classes]))
}
pred = multilayer_perception(x, weights, biases)
cost = tf.reduce_mean(tf.nn.softmax_cross_entropy_with_logits(labels=y, logits=pred))
optimizer = tf.train.AdamOptimizer(learning_rate=learning_rate).minimize(cost)
init = tf.initialize_all_variables()
with tf.Session() as sess:
sess.run(init)
for epoch in range(training_epochs):
avg_cost = 0
total_batch = int(mnist.train.num_examples / batch_size)
for i in range(total_batch):
batch_xs, batch_ys = mnist.train.next_batch(batch_size)
sess.run(optimizer, feed_dict={x: batch_xs, y: batch_ys})
avg_cost += sess.run(cost, feed_dict={x: batch_xs, y: batch_ys}) / total_batch
if epoch % display_step == 0:
print("Epoch:", "%04d" % (epoch+1), "cost=", "{:.9f}".format(avg_cost))
print("Optimization Finished!!!!!!")
correct_prediction = tf.equal(tf.argmax(pred, 1), tf.argmax(y, 1))
accuracy = tf.reduce_mean(tf.cast(correct_prediction, "float"))
print("Accuracy:", accuracy.eval({x: mnist.test.images, y: mnist.test.labels}))
RNN
import tensorflow as tf
from tensorflow.examples.tutorials.mnist import input_data
mnist = input_data.read_data_sets("mnist_data", one_hot=True)
from tensorflow.contrib import rnn
import numpy as np
learning_rate = 0.001
training_iters = 100000
batch_size = 128
display_step = 10
n_input = 28
n_steps = 28
n_hidden = 128
n_classes = 10
x = tf.placeholder("float", [None, n_steps, n_input])
istate = tf.placeholder("float", [None, 2* n_hidden])
y = tf.placeholder("float", [None, n_classes])
weights = {
"hidden": tf.Variable(tf.random_normal([n_input, n_hidden])),
"out": tf.Variable(tf.random_normal([n_hidden, n_classes]))
}
biases = {
"hidden": tf.Variable(tf.random_normal([n_hidden])),
"out": tf.Variable(tf.random_normal([n_classes]))
}
def RNN(_X, _istate, _weights, _biases):
_X = tf.transpose(_X, [1, 0, 2])
_X = tf.reshape(_X, [-1, n_input])
_X = tf.matmul(_X, _weights['hidden']) + _biases['hidden']
lstm_cell = rnn.BasicLSTMCell(n_hidden, state_is_tuple=False)
_X = tf.split(_X, n_steps, 0)
outputs, states = rnn.static_rnn(lstm_cell, _X, initial_state=_istate)
return tf.matmul(outputs[-1], weights['out']) + biases['out']
pred = RNN(x, istate, weights, biases)
cost = tf.reduce_mean(tf.nn.softmax_cross_entropy_with_logits(logits=pred, labels=y))
optimizer = tf.train.AdamOptimizer(learning_rate=learning_rate).minimize(cost)
correct_pred = tf.equal(tf.argmax(pred, 1), tf.argmax(y, 1))
accuracy = tf.reduce_mean(tf.cast(correct_pred, tf.float32))
init = tf.initialize_all_variables()
with tf.Session() as sess:
sess.run(init)
step = 1
while step * batch_size < training_iters:
batch_xs, batch_ys = mnist.train.next_batch(batch_size)
batch_xs = batch_xs.reshape((batch_size, n_steps, n_input))
sess.run(optimizer,feed_dict={x: batch_xs, y: batch_ys,
istate: np.zeros((batch_size, 2 * n_hidden))})
if step % display_step == 0:
acc = sess.run(accuracy, feed_dict={x: batch_xs, y: batch_ys,
istate: np.zeros((batch_size, 2 * n_hidden))})
loss = sess.run(cost, feed_dict={x: batch_xs, y: batch_ys,
istate: np.zeros((batch_size, 2 * n_hidden))})
print("Iter " + str(step * batch_size) + ", Minibatch Loss= " + "{:.6f}".format(loss) +
", Training Accuracy= " + "{:.5f}".format(acc))
step += 1
print("Optimization Finished!!!!!!!")
test_len = 256
test_data = mnist.test.images[:test_len].reshape((-1, n_steps, n_input))
test_label = mnist.test.labels[:test_len]
print("Testing Accuracy:", sess.run(accuracy, feed_dict={x: test_data, y: test_label,
istate: np.zeros((test_len, 2 * n_hidden))}))