激活函数的实现(sigmoid、softmax、tanh、relu、leakyrelu、elu、selu、softplus):https://www.cnblogs.com/xiximayou/p/12713081.html
优化器的实现(SGD、Nesterov、Adagrad、Adadelta、RMSprop、Adam):https://www.cnblogs.com/xiximayou/p/12713594.html
首先是所有的代码:
from __future__ importprint_function, divisionfrom terminaltables importAsciiTableimportnumpy as npimportprogressbarfrom mlfromscratch.utils importbatch_iteratorfrom mlfromscratch.utils.misc importbar_widgetsclassNeuralNetwork():"""Neural Network. Deep Learning base model.
Parameters:
-----------
optimizer: class
The weight optimizer that will be used to tune the weights in order of minimizing
the loss.
loss: class
Loss function used to measure the model's performance. SquareLoss or CrossEntropy.
validation: tuple
A tuple containing validation data and labels (X, y)"""
def __init__(self, optimizer, loss, validation_data=None):
self.optimizer=optimizer
self.layers=[]
self.errors= {"training": [], "validation": []}
self.loss_function=loss()
self.progressbar= progressbar.ProgressBar(widgets=bar_widgets)
self.val_set=Noneifvalidation_data:
X, y=validation_data
self.val_set= {"X": X, "y": y}defset_trainable(self, trainable):"""Method which enables freezing of the weights of the network's layers."""
for layer inself.layers:
layer.trainable=trainabledefadd(self, layer):"""Method which adds a layer to the neural network"""
#If this is not the first layer added then set the input shape
#to the output shape of the last added layer
ifself.layers:
layer.set_input_shape(shape=self.layers[-1].output_shape())#If the layer has weights that needs to be initialized
if hasattr(layer, 'initialize'):
layer.initialize(optimizer=self.optimizer)#Add layer to the network
self.layers.append(layer)deftest_on_batch(self, X, y):"""Evaluates the model over a single batch of samples"""y_pred= self._forward_pass(X, training=False)
loss=np.mean(self.loss_function.loss(y, y_pred))
acc=self.loss_function.acc(y, y_pred)returnloss, accdeftrain_on_batch(self, X, y):"""Single gradient update over one batch of samples"""y_pred=self._forward_pass(X)
loss=np.mean(self.loss_function.loss(y, y_pred))
acc=self.loss_function.acc(y, y_pred)#Calculate the gradient of the loss function wrt y_pred
loss_grad =self.loss_function.gradient(y, y_pred)#Backpropagate. Update weights
self._backward_pass(loss_grad=loss_grad)returnloss, accdeffit(self, X, y, n_epochs, batch_size):"""Trains the model for a fixed number of epochs"""
for _ inself.progressbar(range(n_epochs)):
batch_error=[]for X_batch, y_batch in batch_iterator(X, y, batch_size=batch_size):
loss, _=self.train_on_batch(X_batch, y_batch)
batch_error.append(loss)
self.errors["training"].append(np.mean(batch_error))if self.val_set is notNone:
val_loss, _= self.test_on_batch(self.val_set["X"], self.val_set["y"])
self.errors["validation"].append(val_loss)return self.errors["training"], self.errors["validation"]def _forward_pass(self, X, training=True):"""Calculate the output of the NN"""layer_output=Xfor layer inself.layers:
layer_output=layer.forward_pass(layer_output, training)returnlayer_outputdef_backward_pass(self, loss_grad):"""Propagate the gradient 'backwards' and update the weights in each layer"""
for layer inreversed(self.layers):
loss_grad=layer.backward_pass(loss_grad)def summary(self, name="Model Summary"):#Print model name
print(AsciiTable([[name]]).table)#Network input shape (first layer's input shape)
print ("Input Shape: %s" %str(self.layers[0].input_shape))#Iterate through network and get each layer's configuration
table_data = [["Layer Type", "Parameters", "Output Shape"]]
tot_params=0for layer inself.layers:
layer_name=layer.layer_name()
params=layer.parameters()
out_shape=layer.output_shape()
table_data.append([layer_name, str(params), str(out_shape)])
tot_params+=params#Print network configuration table
print(AsciiTable(table_data).table)print ("Total Parameters: %d\n" %tot_params)defpredict(self, X):"""Use the trained model to predict labels of X"""
return self._forward_pass(X, training=False)
接着我们来一个一个函数进行分析:
1、初始化__init__:这里面定义好优化器optimizer、模型层layers、错误errors、损失函数loss_function、用于显示进度条progressbar,这里从mlfromscratch.utils.misc中导入了bar_widgets,我们看看这是什么:
bar_widgets =['Training:', progressbar.Percentage(), ' ', progressbar.Bar(marker="-", left="[", right="]"),' ', progressbar.ETA()
]
2、set_trainable():用于设置哪些模型层需要进行参数的更新
3、add():将一个模块放入到卷积神经网络中,例如卷积层、池化层、激活层等等。
4、test_on_batch():使用batch进行测试,这里不需要进行反向传播。
5、train_on_batch():使用batch进行训练,包括前向传播计算损失以及反向传播更新参数。
6、fit():喂入数据进行训练或验证,这里需要定义好epochs和batch_size的大小,同时有一个读取数据的函数batch_iterator(),位于mlfromscratch.utils下的data_manipulation.py中:
def batch_iterator(X, y=None, batch_size=64):"""Simple batch generator"""n_samples=X.shape[0]for i innp.arange(0, n_samples, batch_size):
begin, end= i, min(i+batch_size, n_samples)if y is notNone:yieldX[begin:end], y[begin:end]else:yield X[begin:end]
7、_forward_pass():模型层的前向传播。
8、_backward_pass():模型层的反向传播。
9、summary():用于输出模型的每层的类型、参数数量以及输出大小。
10、predict():用于输出预测值。
不难发现,该代码是借鉴了tensorflow中的一些模块的设计思想。