TensorFlow实战 前向传播

导入模块

import os 
import tensorflow as tf
from tensorflow import keras
from tensorflow.keras import datasets,layers,optimizers

数据加载

(x,y),(x_val,y_val) = datasets.mnist.load_data()
x = tf.convert_to_tensor(x,dtype=tf.float32)/255.
y = tf.convert_to_tensor(y,dtype=tf.int32)

创建数据集

train_dataset = tf.data.Dataset.from_tensor_slices((x,y)).batch(200)
train_iter = iter(train_dataset)
sample = next(train_iter)

维度变换

#[b,784] => [b,512] => [b,256] => [b,10]
w1 = tf.random.truncated_normal([784,512])
b1 = tf.zeros([512])
w2 = tf.random.truncated_normal([512,256])
b2 = tf.zeros([256])
w3 = tf.random.truncated_normal([256,10])
b3 = tf.zeros([10])
for (x,y) in train_dataset:
    x = tf.reshape(x,[-1,28*28])
    #h = x@w + b
    #[b,784]@[784,512]+[512] => [b,512]+[512] => [b,512]+[b,512]
    h1 = x@w1 + b1
    h1 = tf.nn.relu(h1) #非线性转化
    h2 = h1@w2 + b2
    h2 = tf.nn.relu(h2) #非线性转化
    out = h2@w3 + b3

计算误差

y_onehot = tf.one_hot(y,depth=10)
        loss = tf.square(y_onehot-out)
        loss = tf.reduce_mean(loss)

计算梯度

 with tf.GradientTape() as tape:
            #h = x@w + b
            #[b,784]@[784,512]+[512] => [b,512]+[512] => [b,512]+[b,512]
            h1 = x@w1 + b1
            h1 = tf.nn.relu(h1) #非线性转化
            h2 = h1@w2 + b2
            h2 = tf.nn.relu(h2) 
            out = h2@w3 + b3

            y_onehot = tf.one_hot(y,depth=10)
            loss = tf.square(y_onehot-out)
            loss = tf.reduce_mean(loss)
        grads = tape.gradient(loss,[w1,b1,w2,b2,w3,b3])

更新权值和偏置项

b1.assign_sub(lr*grads[1])
w2.assign_sub(lr*grads[2])
b2.assign_sub(lr*grads[3])
w3.assign_sub(lr*grads[4])
b3.assign_sub(lr*grads[5])

完整代码

import os
from random import sample 
import tensorflow as tf
from tensorflow import keras
from tensorflow.keras import datasets,layers,optimizers

os.environ['TF_CPP_MIN_LOG_LEVEL']='2'
#数据加载
(x,y),(x_val,y_val) = datasets.mnist.load_data()
x = tf.convert_to_tensor(x,dtype=tf.float32)/255.
y = tf.convert_to_tensor(y,dtype=tf.int32)
print(tf.reduce_max(x),tf.reduce_min(x))
print(tf.reduce_max(y),tf.reduce_min(y))
train_dataset = tf.data.Dataset.from_tensor_slices((x,y)).batch(200)
train_iter = iter(train_dataset)
sample = next(train_iter)

#[b,784] => [b,512] => [b,256] => [b,10]
w1 = tf.Variable(tf.random.truncated_normal([784,512],stddev=0.1))
b1 = tf.Variable(tf.zeros([512]))
w2 = tf.Variable(tf.random.truncated_normal([512,256],stddev=0.1))
b2 = tf.Variable(tf.zeros([256]))
w3 = tf.Variable(tf.random.truncated_normal([256,10],stddev=0.1))
b3 = tf.Variable(tf.zeros([10]))
lr = 0.001
for epoch in range(30):
    for step,(x,y) in enumerate(train_dataset):
        x = tf.reshape(x,[-1,28*28])#维度变换
        with tf.GradientTape() as tape:
            #h = x@w + b
            #[b,784]@[784,512]+[512] => [b,512]+[512] => [b,512]+[b,512]
            h1 = x@w1 + b1
            h1 = tf.nn.relu(h1) #非线性转化
            h2 = h1@w2 + b2
            h2 = tf.nn.relu(h2) 
            out = h2@w3 + b3

            y_onehot = tf.one_hot(y,depth=10)
            loss = tf.square(y_onehot-out)
            loss = tf.reduce_mean(loss)
        grads = tape.gradient(loss,[w1,b1,w2,b2,w3,b3])
        #w = w-lr*grads
        w1.assign_sub(lr*grads[0])
        b1.assign_sub(lr*grads[1])
        w2.assign_sub(lr*grads[2])
        b2.assign_sub(lr*grads[3])
        w3.assign_sub(lr*grads[4])
        b3.assign_sub(lr*grads[5])
        if step%100==0:
            print(epoch,loss.numpy())
  • 0
    点赞
  • 0
    收藏
    觉得还不错? 一键收藏
  • 0
    评论
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值