pytorch神经网络入门(九)

一、概述

用训练好的模型进行nimist数据集分类,并且可视化展示

二、代码

import torch
import torch.nn as nn
from tensorboardX import SummaryWriter
from pickle import FALSE
from operator import truediv
import torchvision
import torch.utils.data as Data
from sklearn.metrics import accuracy_score
import torchvision.utils as vutils
from matplotlib import pyplot as plt

logger = SummaryWriter(log_dir="data/log")
 
class ConvNet(nn.Module):
    def __init__(self):
         super(ConvNet, self).__init__()
 
         self.conv1 = nn.Sequential(
             nn.Conv2d(1, 16, 3, 1, 1),
             nn.ReLU(),
             nn.AvgPool2d(2, 2)
         )
 
         self.conv2 = nn.Sequential(
             nn.Conv2d(16, 32, 3, 1, 1),
             nn.ReLU(),
             nn.MaxPool2d(2, 2)
         )
 
         self.fc = nn.Sequential(
             nn.Linear(32 * 7 * 7, 128),
             nn.ReLU(),
             nn.Linear(128, 64),
             nn.ReLU()
         )
 
         self.out = nn.Linear(64, 10)
 
    def forward(self, x):
         x = self.conv1(x)
         x = self.conv2(x)
         x = x.view(x.size(0), -1)
         x = self.fc(x)
         output = self.out(x)
         return output

#数据下载
# 准备训练用的MNIST数据集
train_data = torchvision.datasets.MNIST(
     root = "./data/MNIST",  # 提取数据的路径
     train=True, # 使用MNIST内的训练数据
     transform=torchvision.transforms.ToTensor(),    # 转换成torch.tensor
     download=FALSE   # 如果是第一次运行的话,置为True,表示下载数据集到root目录
)
 
 # 定义loader
train_loader = Data.DataLoader(
     dataset=train_data,
     batch_size=128,
     shuffle=True,
     num_workers=0
 )
 
test_data = torchvision.datasets.MNIST(
     root="./dataset/MNIST",
     train=False,    # 使用测试数据
     download=False
 )
 # 将测试数据压缩到0-1
test_data_x = test_data.data.type(torch.FloatTensor) / 255.0
test_data_x = torch.unsqueeze(test_data_x, dim=1)
test_data_y = test_data.targets


def training(istrain):
    if istrain==0:
        MyConvNet = ConvNet()
        print(MyConvNet)
        # 获取优化器和损失函数
        optimizer = torch.optim.Adam(MyConvNet.parameters(), lr=3e-4)
        loss_func = nn.CrossEntropyLoss()
        log_step_interval = 100  # 记录的步数间隔

        for epoch in range(2):
            print("epoch:", epoch)
            # 每一轮都遍历一遍数据加载器
            for step, (x, y) in enumerate(train_loader):
                # 前向计算->计算损失函数->(从损失函数)反向传播->更新网络
                predict = MyConvNet(x)
                loss = loss_func(predict, y)
                optimizer.zero_grad()  # 清空梯度(可以不写)
                loss.backward()  # 反向传播计算梯度
                optimizer.step()  # 更新网络
                global_iter_num = epoch * len(train_loader) + step + 1  # 计算当前是从训练开始时的第几步(全局迭代次数)
                if global_iter_num % log_step_interval == 0:
                    # 控制台输出一下
                    print("global_step:{}, loss:{:.2}".format(global_iter_num, loss.item()))
                    # 添加的第一条日志:损失函数-全局迭代次数
                    logger.add_scalar("train loss", loss.item(), global_step=global_iter_num)
                    # 在测试集上预测并计算正确率
                    test_predict = MyConvNet(test_data_x)
                    _, predict_idx = torch.max(test_predict, 1)  # 计算softmax后的最大值的索引,即预测结果
                    acc = accuracy_score(test_data_y, predict_idx)
                    # 添加第二条日志:正确率-全局迭代次数
                    logger.add_scalar("test accuary", acc.item(), global_step=global_iter_num)
                    # 添加第三条日志:这个batch下的128张图像
                    img = vutils.make_grid(x, nrow=12)
                    logger.add_image("train image sample", img, global_step=global_iter_num)
                    # 添加第三条日志:网络中的参数分布直方图
                    for name, param in MyConvNet.named_parameters():
                        logger.add_histogram(name, param.data.numpy(), global_step=global_iter_num)
                torch.save(MyConvNet.state_dict(), 'pth/cnn_picture_modelparam_%d.pt' % (epoch))
                torch.save(MyConvNet, "pth/model_%d.pth" % epoch)

                # vutils.save_image==>将给定的张量保存为图像
    else:
        pass

def usetrainedmodel(istrain):
    if istrain==0:
        MyConvNet = torch.load("pth/model_1.pth")
        # model = ConvNet()
        MyConvNet.load_state_dict(torch.load("pth/cnn_picture_modelparam_1.pt"))

        # 获取优化器和损失函数
        optimizer = torch.optim.Adam(MyConvNet.parameters(), lr=3e-4)
        loss_func = nn.CrossEntropyLoss()
        log_step_interval = 100  # 记录的步数间隔

        for epoch in range(2):
            print("epoch:", epoch)
            # 每一轮都遍历一遍数据加载器
            for step, (x, y) in enumerate(train_loader):
                # 前向计算->计算损失函数->(从损失函数)反向传播->更新网络
                predict = MyConvNet(x)

                loss = loss_func(predict, y)
                optimizer.zero_grad()  # 清空梯度(可以不写)
                loss.backward()  # 反向传播计算梯度
                optimizer.step()  # 更新网络
                global_iter_num = epoch * len(train_loader) + step + 1  # 计算当前是从训练开始时的第几步(全局迭代次数)
                if global_iter_num % log_step_interval == 0:
                    # 控制台输出一下
                    print("global_step:{}, loss:{:.2}".format(global_iter_num, loss.item()))
                    # 添加的第一条日志:损失函数-全局迭代次数
                    logger.add_scalar("train loss", loss.item(), global_step=global_iter_num)
                    # 在测试集上预测并计算正确率
                    test_predict = MyConvNet(test_data_x)
                    _, predict_idx = torch.max(test_predict, 1)  # 计算softmax后的最大值的索引,即预测结果
                    acc = accuracy_score(test_data_y, predict_idx)
                    # 添加第二条日志:正确率-全局迭代次数
                    logger.add_scalar("test accuary", acc.item(), global_step=global_iter_num)
                    # 添加第三条日志:这个batch下的128张图像
                    img = vutils.make_grid(x, nrow=12)
                    logger.add_image("train image sample", img, global_step=global_iter_num)
                    # 添加第三条日志:网络中的参数分布直方图
                    for name, param in MyConvNet.named_parameters():
                        logger.add_histogram(name, param.data.numpy(), global_step=global_iter_num)

    else:
        pass

def visibletrainresult():
    MyConvNet = torch.load("pth/model_1.pth")
    # model = ConvNet()
    MyConvNet.load_state_dict(torch.load("pth/cnn_picture_modelparam_1.pt"))

    for step,(x, y) in enumerate(train_loader):

        print(x.size())#torch.Size([128, 1, 28, 28])
        print(y.size())#[128]
        predict = MyConvNet(x)
        print("predict:")
        print(predict)
        pred = predict.argmax(dim=1)
        plot_image(x, pred, 'test')


def plot_image(img, label, name):

    fig = plt.figure()
    for i in range(6):
        plt.subplot(2, 3, i + 1)
        plt.tight_layout()
        plt.imshow(img[i][0]*0.3081+0.1307, cmap='gray', interpolation='none')
        plt.title("{}: {}".format(name, label[i].item()))
        plt.xticks([])
        plt.yticks([])
    plt.show()

if __name__=='__main__':
    training(1)
    usetrainedmodel(1)
    visibletrainresult()

三、附图 

 四、总结

注意区别两种获取图片数据方式

  for step,(x, y) in enumerate(train_loader):
        print(step)
        print(x.size())#torch.Size([128, 1, 28, 28])
        print(y.size())#[128]
 x, y = next(iter(test_loader))

  • 0
    点赞
  • 1
    收藏
    觉得还不错? 一键收藏
  • 0
    评论

“相关推荐”对你有帮助么?

  • 非常没帮助
  • 没帮助
  • 一般
  • 有帮助
  • 非常有帮助
提交
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值