python bp神经网络 异或_基于python的BP神经网络及异或实现过程解析

本文介绍了一个简单的BP神经网络实现,用于模拟非线性函数。文章详细展示了网络的初始化、训练过程,包括权重和偏置的确定、梯度下降优化以及损失函数的计算。在训练过程中,通过迭代更新参数以求得收敛。最后,通过示例数据进行预测,并绘制了损失函数随迭代次数的变化曲线,以展示网络的收敛效果。
摘要由CSDN通过智能技术生成

BP神经网络是最简单的神经网络模型了,三层能够模拟非线性函数效果。

难点:

如何确定初始化参数?

如何确定隐含层节点数量?

迭代多少次?如何更快收敛?

如何获得全局最优解?

'''

neural networks

created on 2019.9.24

author: vince

'''

import math

import logging

import numpy

import random

import matplotlib.pyplot as plt

'''

neural network

'''

class NeuralNetwork:

def __init__(self, layer_nums, iter_num = 10000, batch_size = 1):

self.__ILI = 0;

self.__HLI = 1;

self.__OLI = 2;

self.__TLN = 3;

if len(layer_nums) != self.__TLN:

raise Exception("layer_nums length must be 3");

self.__layer_nums = layer_nums; #array [layer0_num, layer1_num ...layerN_num]

self.__iter_num = iter_num;

self.__batch_size = batch_size;

def train(self, X, Y):

X = numpy.array(X);

Y = numpy.array(Y);

self.L = [];

#initialize parameters

self.__weight = [];

self.__bias = [];

self.__step_len = [];

for layer_index in range(1, self.__TLN):

self.__weight.append(numpy.random.rand(self.__layer_nums[layer_index - 1], self.__layer_nums[layer_index]) * 2 - 1.0);

self.__bias.append(numpy.random.rand(self.__layer_nums[layer_index]) * 2 - 1.0);

self.__step_len.append(0.3);

logging.info("bias:%s" % (self.__bias));

logging.info("weight:%s" % (self.__weight));

for iter_index in range(self.__iter_num):

sample_index = random.randint(0, len(X) - 1);

logging.debug("-----round:%s, select sample %s-----" % (iter_index, sample_index));

output = self.forward_pass(X[sample_index]);

g = (-output[2] + Y[sample_index]) * self.activation_drive(output[2]);

logging.debug("g:%s" % (g));

for j in range(len(output[1])):

self.__weight[1][j] += self.__step_len[1] * g * output[1][j];

self.__bias[1] -= self.__step_len[1] * g;

e = [];

for i in range(self.__layer_nums[self.__HLI]):

e.append(numpy.dot(g, self.__weight[1][i]) * self.activation_drive(output[1][i]));

e = numpy.array(e);

logging.debug("e:%s" % (e));

for j in range(len(output[0])):

self.__weight[0][j] += self.__step_len[0] * e * output[0][j];

self.__bias[0] -= self.__step_len[0] * e;

l = 0;

for i in range(len(X)):

predictions = self.forward_pass(X[i])[2];

l += 0.5 * numpy.sum((predictions - Y[i]) ** 2);

l /= len(X);

self.L.append(l);

logging.debug("bias:%s" % (self.__bias));

logging.debug("weight:%s" % (self.__weight));

logging.debug("loss:%s" % (l));

logging.info("bias:%s" % (self.__bias));

logging.info("weight:%s" % (self.__weight));

logging.info("L:%s" % (self.L));

def activation(self, z):

return (1.0 / (1.0 + numpy.exp(-z)));

def activation_drive(self, y):

return y * (1.0 - y);

def forward_pass(self, x):

data = numpy.copy(x);

result = [];

result.append(data);

for layer_index in range(self.__TLN - 1):

data = self.activation(numpy.dot(data, self.__weight[layer_index]) - self.__bias[layer_index]);

result.append(data);

return numpy.array(result);

def predict(self, x):

return self.forward_pass(x)[self.__OLI];

def main():

logging.basicConfig(level = logging.INFO,

format = '%(asctime)s %(filename)s[line:%(lineno)d] %(levelname)s %(message)s',

datefmt = '%a, %d %b %Y %H:%M:%S');

logging.info("trainning begin.");

nn = NeuralNetwork([2, 2, 1]);

X = numpy.array([[0, 0], [1, 0], [1, 1], [0, 1]]);

Y = numpy.array([0, 1, 0, 1]);

nn.train(X, Y);

logging.info("trainning end. predict begin.");

for x in X:

print(x, nn.predict(x));

plt.plot(nn.L)

plt.show();

if __name__ == "__main__":

main();

具体收敛效果

以上就是本文的全部内容,希望对大家的学习有所帮助,也希望大家多多支持爱蒂网。

好的,下面是 Python 实现的 BP 神经网络解决异或问题的代码: ```python import numpy as np # 定义 sigmoid 函数 def sigmoid(x): return 1 / (1 + np.exp(-x)) # 定义 sigmoid 函数的导数 def sigmoid_derivative(x): return x * (1 - x) # 训练 BP 神经网络 def train(X, y, hidden_neurons=3, epochs=10000, alpha=0.1): # 初始化输入层权重和偏置 input_neurons = X.shape[1] w1 = np.random.uniform(size=(input_neurons, hidden_neurons)) b1 = np.random.uniform(size=(1, hidden_neurons)) # 初始化输出层权重和偏置 output_neurons = y.shape[1] w2 = np.random.uniform(size=(hidden_neurons, output_neurons)) b2 = np.random.uniform(size=(1, output_neurons)) # 开始训练 for i in range(epochs): # 前向传播 hidden_layer_input = np.dot(X, w1) + b1 hidden_layer_output = sigmoid(hidden_layer_input) output_layer_input = np.dot(hidden_layer_output, w2) + b2 output_layer_output = sigmoid(output_layer_input) # 反向传播 error = y - output_layer_output d_output = error * sigmoid_derivative(output_layer_output) error_hidden_layer = d_output.dot(w2.T) d_hidden_layer = error_hidden_layer * sigmoid_derivative(hidden_layer_output) # 更新权重和偏置 w2 += hidden_layer_output.T.dot(d_output) * alpha b2 += np.sum(d_output, axis=0, keepdims=True) * alpha w1 += X.T.dot(d_hidden_layer) * alpha b1 += np.sum(d_hidden_layer, axis=0, keepdims=True) * alpha return w1, b1, w2, b2 # 测试 BP 神经网络 def predict(X, w1, b1, w2, b2): # 前向传播 hidden_layer_input = np.dot(X, w1) + b1 hidden_layer_output = sigmoid(hidden_layer_input) output_layer_input = np.dot(hidden_layer_output, w2) + b2 output_layer_output = sigmoid(output_layer_input) return output_layer_output # 定义输入和输出数据 X = np.array([[0, 0], [0, 1], [1, 0], [1, 1]]) y = np.array([[0], [1], [1], [0]]) # 训练 BP 神经网络 w1, b1, w2, b2 = train(X, y, hidden_neurons=3, epochs=10000, alpha=0.1) # 测试 BP 神经网络 predictions = predict(X, w1, b1, w2, b2) print(predictions) ``` 代码中使用了 sigmoid 函数作为激活函数,并且使用了 sigmoid 函数的导数来计算反向传播的梯度。在训练时,通过随机初始化权重和偏置,使用前向传播和反向传播来更新权重和偏置,最终得到训练好的 BP 神经网络。在测试时,使用训练好的 BP 神经网络来预测输出结果。
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值