BP神经网络实战
前段时间看了BP神经网络,并进行回归预测,下面从三种方法进行阐述。
方法一、直接使用波斯顿房价预测案例进行简单修改,话不多说,源码如下:(代码备注很清晰,一看既懂)
from sklearn.metrics import mean_squared_error
from sklearn.metrics import mean_absolute_error
from sklearn.metrics import r2_score
import tensorflow as tf
import matplotlib.pyplot as plt
import numpy as np
import pandas as pd
from sklearn.utils import shuffle
train_epochs = 1
learning_rate = 0.01
def get_data():
mydf = pd.read_csv('temp.csv', sep='\s+',encoding='utf-8')
print(mydf)
df = mydf.values
print(df)
return df
def normal(df):
for i in range(10):
df[:, i] = (df[:, i] - df[:, i].min()) / (df[:, i].max() - df[:, i].min())
x_data = df[:, :10]
y_data = df[:, 10]
return x_data, y_data
def model():
x = tf.placeholder(tf.float32, [None, 10], name="X")
y = tf.placeholder(tf.float32, [None, 1], name="Y")
with tf.name_scope("Model"):
w = tf.Variable(tf.random_normal([10, 1], stddev=0.01), name="W")
b = tf.Variable(1.0, name="b")
def model(x, w, b):
return tf.matmul(x, w) + b
pred = model(x, w, b)
print("y_pred is %s"%(pred))
with tf.name_scope("LossFunction"):
loss_function = tf.reduce_mean(tf.pow(y - pred, 2))
print("loss_function is ")
print(loss_function)
optimizer = tf.train.GradientDescentOptimizer(learning_rate).minimize(loss_function)
return loss_function, optimizer, pred, x, y, b, w
def run_model(loss_function, optimizer, pred, x_data, y_data, x, y, b, w):
sess = tf.Session()
init = tf.global_variables_initializer()
logdir = './test'
sum_loss_op = tf.summary.scalar("loss", loss_function)
merged = tf.summary.merge_all()
sess.run(init)
writer = tf.summary.FileWriter(logdir, sess.graph)
loss_list = []
b0 = []
Add_train_y = []
Add_predict_y = []
for epoch in range(train_epochs):
loss_sum = 0.0
for xs, ys in zip(x_data, y_data):
xs = xs.reshape(1, 10)
ys = ys.reshape(1, 1)
_, summary_str, loss = sess.run([optimizer, sum_loss_op, loss_function], feed_dict={x: xs, y: ys})
writer.add_summary(summary_str, epoch)
loss_sum = loss_sum + loss
x_data, y_data = shuffle(x_data, y_data)
b0temp = b.eval(session=sess)
w0temp = w.eval(session=sess)
loss_average = loss_sum / len(y_data)
b0.append(b0temp)
loss_list.append(loss_average)
print("epoch=", epoch + 1, "loss=", loss_average, "b=", b0temp, "w=", w0temp)
Add_train_y.append(np.sqrt(mean_squared_error(y_data, sess.run(pred, feed_dict={x: x_data}))))
print("Add_train_y is %s"%(Add_train_y))
return loss_list, pred, Add_train_y, sess
def plot_all(loss_list, x_data, y_data, pred, Add_train_y, sess, x, y):
plt.plot(loss_list)
n = 1
print(n+2)
x_test = x_data[n]
print("x_test is %s"%(x_test))
x_test = x_test.reshape(1, 10)
print(x_test)
predict = sess.run(pred, feed_dict={x: x_test})
print("预测值:%f" % predict)
target = y_data[n]
print("标签值:%f" % target)
plt.show()
test_predictions = sess.run(pred, feed_dict={x: x_data})
plt.scatter(y_data, test_predictions)
plt.xlabel('True Values ')
plt.ylabel('Predictions ')
plt.axis('equal')
plt.xlim(plt.xlim())
plt.ylim(plt.ylim())
_ = plt.plot([-100, 100], [-100, 100])
plt.show()
plt.figure()
plt.xlabel('Epoch')
plt.ylabel('Mean Abs Error ')
plt.plot(np.arange(10), Add_train_y,
label = 'MAE')
plt.legend()
plt.ylim([0, 10])
plt.show()
def main():
df = get_data()
print(df)
x_data, y_data = normal(df)
print("X_data is %s,y_data is %s"%(x_data,y_data))
loss_function, optimizer, pred, x, y, b, w = model()
loss_list, pred, Add_train_y, sess = run_model(loss_function, optimizer, pred, x_data, y_data, x, y, b, w)
plot_all(loss_list, x_data, y_data, pred, Add_train_y, sess, x, y)
if __name__ == '__main__':
main()
实验结果如下:
![在这里插入图片描述](https://img-blog.csdnimg.cn/20200520172021138.png?x-oss-process=image/watermark,type_ZmFuZ3poZW5naGVpdGk,shadow_10,text_aHR0cHM6Ly9ibG9nLmNzZG4ubmV0L3poYW5neGluZzY2NjY=,size_16,color_FFFFFF,t_70#pic_center)
方法二、tensorflow结合正则。
import tensorflow as tf
import pandas as pd
import numpy as np
createVar = locals()
'''
建立一个网络结构可变的BP神经网络通用代码:
在训练时各个参数的意义:
hidden_floors_num:隐藏层的个数
every_hidden_floor_num:每层隐藏层的神经元个数
learning_rate:学习速率
activation:激活函数
regularization:正则化方式
regularization_rate:正则化比率
total_step:总的训练次数
train_data_path:训练数据路径
model_save_path:模型保存路径
利用训练好的模型对验证集进行验证时各个参数的意义:
model_save_path:模型保存路径
validate_data_path:验证集路径
precision:精度
利用训练好的模型进行预测时各个参数的意义:
model_save_path:模型的保存路径
predict_data_path:预测数据路径
predict_result_save_path:预测结果保存路径
'''
hidden_floors_num = 1
every_hidden_floor_num = [50]
learning_rate = 0.00001
activation = 'tanh'
regularization = 'L2'
regularization_rate = 0.7
total_step = 100000
train_data_path = './sum.csv'
model_save_path = './model/predict_model'
precision = 0.5
model_save_path = './model/predict_model'
predict_data_path = './data2.csv'
predict_result_save_path = './result.csv'
def inputs(train_data_path):
train_data = pd.read_csv(train_data_path)
print(train_data)
X = np.array(train_data.iloc[:, :-1])
Y = np.array(train_data.iloc[:, -1:])
return X, Y
def make_hidden_layer(pre_lay_num, cur_lay_num, floor):
createVar['w' + str(floor)] = tf.Variable(tf.random_normal([pre_lay_num, cur_lay_num], stddev=1))
createVar['b' + str(floor)] = tf.Variable(tf.random_normal([cur_lay_num], stddev=1))
return eval('w'+str(floor)), eval('b'+str(floor))
def initial_w_and_b(all_floors_num):
for floor in range(2, hidden_floors_num+3):
pre_lay_num = all_floors_num[floor-2]
cur_lay_num = all_floors_num[floor-1]
w_floor, b_floor = make_hidden_layer(pre_lay_num, cur_lay_num, floor)
createVar['w' + str(floor)] = w_floor
createVar['b' + str(floor)] = b_floor
def cal_floor_output(x, floor):
w_floor = eval('w'+str(floor))
b_floor = eval('b'+str(floor))
if activation == 'sigmoid':
output = tf.sigmoid(tf.matmul(x, w_floor) + b_floor)
if activation == 'tanh':
output = tf.tanh(tf.matmul(x, w_floor) + b_floor)
if activation == 'relu':
output = tf.nn.relu(tf.matmul(x, w_floor) + b_floor)
return output
def inference(x):
output = x
for floor in range(2, hidden_floors_num+2):
output = cal_floor_output(output, floor)
floor = hidden_floors_num+2
w_floor = eval('w'+str(floor))
b_floor = eval('b'+str(floor))
output = tf.matmul(output, w_floor) + b_floor
return output
def loss(x, y_real):
y_pre = inference(x)
if regularization == 'None':
total_loss = tf.reduce_sum(tf.squared_difference(y_real, y_pre))
if regularization == 'L1':
total_loss = 0
for floor in range(2, hidden_floors_num + 3):
w_floor = eval('w' + str(floor))
total_loss = total_loss + tf.contrib.layers.l1_regularizer(regularization_rate)(w_floor)
total_loss = total_loss + tf.reduce_sum(tf.squared_difference(tf.clip_by_value(y_real,1e-8,tf.reduce_max(y_real)), tf.clip_by_value(y_pre,1e-8,tf.reduce_max(y_pre))))
if regularization == 'L2':
total_loss = 0
for floor in range(2, hidden_floors_num + 3):
w_floor = eval('w' + str(floor))
total_loss = total_loss + tf.contrib.layers.l2_regularizer(regularization_rate)(w_floor)
total_loss = total_loss + tf.reduce_sum(tf.squared_difference(tf.clip_by_value(y_real,1e-8,tf.reduce_max(y_real)), tf.clip_by_value(y_pre,1e-8,tf.reduce_max(y_pre))))
return total_loss
def train(total_loss):
train_op = tf.train.GradientDescentOptimizer(learning_rate).minimize(total_loss)
return train_op
def train_model(hidden_floors_num, every_hidden_floor_num, learning_rate, activation, regularization,
regularization_rate, total_step, train_data_path, model_save_path):
X, Y = inputs(train_data_path)
print("X is %s,Y is %s"%(X,Y))
print("X is %s"%(X))
X_dim = X.shape[1]
print("X_dim is %s"%(X_dim))
all_floors_num = [X_dim] + every_hidden_floor_num + [1]
temp = model_save_path.split('/')
model_name = temp[-1]
parameter_path = ''
for i in range(len(temp)-1):
parameter_path = parameter_path + temp[i] + '/'
parameter_path = parameter_path + model_name + '_parameter.txt'
with open(parameter_path, 'w') as f:
f.write("all_floors_num:")
for i in all_floors_num:
f.write(str(i) + ' ')
f.write('\n')
f.write('activation:')
f.write(str(activation))
x = tf.placeholder(dtype=tf.float32, shape=[None, X_dim])
y_real = tf.placeholder(dtype=tf.float32, shape=[None, 1])
initial_w_and_b(all_floors_num)
y_pre = inference(x)
print("x is %s,y_pre is %s,y_real is %s"%(x,y_pre,y_real))
total_loss = loss(x, y_real)
print("total-losss is %s"%(total_loss))
train_op = train(total_loss)
print("train_op is %s"%(train_op))
train_accuracy = tf.reduce_mean(tf.cast(tf.abs(y_pre - y_real) < precision, tf.float32))
saver = tf.train.Saver()
sess = tf.Session()
init = tf.global_variables_initializer()
sess.run(init)
for step in range(total_step):
sess.run([train_op], feed_dict={x: X[0:, :], y_real: Y[0:, :]})
if step % 1000 == 0:
saver.save(sess, model_save_path)
total_loss_value = sess.run(total_loss, feed_dict={x: X[0:, :], y_real: Y[0:, :]})
print('train step is ', step, ', total loss value is ', total_loss_value,
', train_accuracy', sess.run(train_accuracy, feed_dict={x: X, y_real: Y}),
', precision is ', precision)
saver.save(sess, model_save_path)
print("训练已经完成")
sess.close()
def predict(model_save_path, predict_data_path, predict_result_save_path):
temp = model_save_path.split('/')
model_name = temp[-1]
parameter_path = ''
for i in range(len(temp)-1):
parameter_path = parameter_path + temp[i] + '/'
parameter_path = parameter_path + model_name + '_parameter.txt'
with open(parameter_path, 'r') as f:
lines = f.readlines()
temp = lines[0].split(':')[-1].split(' ')
all_floors_num = []
for i in range(len(temp)-1):
all_floors_num = all_floors_num + [int(temp[i])]
activation = lines[1].split(':')[-1]
hidden_floors_num = len(all_floors_num) - 2
predict_data = pd.read_csv(predict_data_path)
X = np.array(predict_data.iloc[:, :])
X_dim = X.shape[1]
x = tf.placeholder(dtype=tf.float32, shape=[None, X_dim])
initial_w_and_b(all_floors_num)
y_pre = inference(x)
sess = tf.Session()
saver = tf.train.Saver()
with tf.Session() as sess:
try:
saver.restore(sess, model_save_path)
print('模型载入成功!')
except:
print('模型不存在,请先训练模型!')
return
y_pre_value = sess.run(y_pre, feed_dict={x: X[0:, :]})
predict_data_columns = list(predict_data.columns) + ['predict']
data = np.column_stack([X, y_pre_value])
result = pd.DataFrame(data, columns=predict_data_columns)
result.to_csv(predict_result_save_path, index=False)
print('预测结果保存在:', predict_result_save_path)
if __name__ == '__main__':
mode = 'train'
if mode == 'train':
train_model(hidden_floors_num, every_hidden_floor_num, learning_rate, activation, regularization,
regularization_rate, total_step, train_data_path, model_save_path)
if mode == 'predict':
predict(model_save_path, predict_data_path, predict_result_save_path)
实验结果
![在这里插入图片描述](https://img-blog.csdnimg.cn/20200520172621519.png?x-oss-process=image/watermark,type_ZmFuZ3poZW5naGVpdGk,shadow_10,text_aHR0cHM6Ly9ibG9nLmNzZG4ubmV0L3poYW5neGluZzY2NjY=,size_16,color_FFFFFF,t_70#pic_center)
torch实战
dataset.py
from torch.utils.data import Dataset,DataLoader
import pandas as pd
import numpy as np
import torch
class MyDataset(Dataset):
def __init__(self,data_dir):
df = pd.read_csv(data_dir,delim_whitespace=True)
datas = df.to_numpy()
for i in range(10):
col = datas[:,i]
col_max = np.max(col)
col_min = np.min(col)
datas[:, i] = (col-col_min)/(col_max-col_min)
self.xs = torch.from_numpy( np.array(datas[:,:10],dtype=np.float32))
self.ys = torch.from_numpy( np.array(datas[:,10:],dtype=np.float32))
def __len__(self):
return len(self.ys)
def __getitem__(self, item):
return self.xs[item],self.ys[item]
mydataset = MyDataset('temp.csv')
train_Dataloder = DataLoader(dataset=mydataset,batch_size=2000)
if __name__ == '__main__':
for x,y in mydataset:
print(y)
train.py
from torch import nn
import torch
from torch.nn import functional as F
from faceRecognize.dataset import *
from torch import optim
from regress.dataset import train_Dataloder
class Net(nn.Module):
def __init__(self):
super(Net, self).__init__()
self.feature = nn.Sequential(
nn.Linear(in_features=10,out_features=32),
nn.BatchNorm1d(32),
nn.LeakyReLU(),
nn.Linear(in_features=32, out_features=64),
nn.BatchNorm1d(64),
nn.LeakyReLU(),
nn.Linear(in_features=64, out_features=128),
nn.BatchNorm1d(128),
nn.LeakyReLU()
)
self.output = nn.Linear(128,1)
def forward(self, x):
feature = self.feature(x)
output = self.output(feature)
return output
if __name__ == '__main__':
net = Net().cuda()
loss_func = nn.MSELoss()
opt = optim.Adam(net.parameters(),lr=0.0001)
for i in range(1000):
for index,(xs,ys) in enumerate(train_Dataloder):
xs = xs.cuda()
ys = ys.cuda()
output = net(xs)
loss = loss_func(output,ys)
opt.zero_grad()
loss.backward()
opt.step()
print(loss)
需要数据集可以邮箱我:
Github