二、其他---Pytorch深度学习入门

一、二

参考:

PyTorch深度学习快速入门教程(绝对通俗易懂!)【小土堆】笔记_电信保温杯的博客-CSDN博客_pytorct深度学习程序详解

P2. Python编辑器的选择、安装及配置(PyCharm、Jupyter安装)【PyTorch教程】_哔哩哔哩_bilibili

三、TensorBoard的使用

1.TensorBoard介绍

TensorBoard是一个可视化工具,它可以用来展示网络图、张量的指标变化、张量的分布情况等。特别是在训练网络的时候,我们可以设置不同的参数(比如:权重W、偏置B、卷积层数、全连接层数等),使用TensorBoader可以很直观的帮我们进行参数的选择。

2.显示图标和图片

from torch.utils.tensorboard import SummaryWriter
import numpy as np
from PIL import Image

writer = SummaryWriter("logs")
image_path = "data/train/ants_image/6240329_72c01e663e.jpg"
img_PIL = Image.open(image_path)
img_array = np.array(img_PIL)
print(type(img_array))
print(img_array.shape)

writer.add_image("train", img_array, 1, dataformats='HWC')  //显示图
# y = 2x
for i in range(100):
    writer.add_scalar("y=2x", 3*i, i)    //显示数据表

writer.close()

在pycharm命令行输入 

tensorboard --logdir=logs  --port=6007

步骤

  1. 创建SummaryWriter类,传入logs文件输出地址
  2. 图片文件地址
  3. Image类打开图片文件
  4. 将Image类转换numpy格式
  5. SummaryWriter类加载numpy格式图片
  6. 画图
  7. 关闭SummaryWriter流

四、Transforms的使用

transforms主要用于图片的变换,是一个torchvision下的一个工具箱,用于格式转化,视觉处理工具,不用于文本。

具体点是这样

 例如

from PIL import Image
from torchvision import transforms

img_path="dataset/train/ants/0013035.jpg"
img=Image.open(img_path)

tensor_trans=transforms.ToTensor()
tensor_img=tensor_trans(img)

print(tensor_img)

显示ToTensor图片

from PIL import Image
from torch.utils.tensorboard import SummaryWriter
from torchvision import transforms

img_path="dataset/train/ants/0013035.jpg"
img=Image.open(img_path)

writer =SummaryWriter("logs")

tensor_trans=transforms.ToTensor()
tensor_img=tensor_trans(img)

writer.add_image("tensor_img",tensor_img)

print(tensor_img)

Transforms里面有多个工具类:
Compose类:将几个变换组合在一起。
ToTensor类:顾名思义,讲其他格式文件转换成tensor类的格式。
ToPILImage类:转换成Image类。
Nomalize类:用于正则化。
Resize类:裁剪。
RandomCrop:随机裁剪。

1.Tensor的操作

微信图片_20210723234005

五、常见的Transforms

1._ _call()_ _ 方法

class Person:
    def __call__(self,name):
      print("__call()__"+" "+name)

    def hello(self,name):
      print("hello"+" "+name)

person=Person()
person("zhangsan")  #对象调用_ _call()_ _方法不用点
person.hello("hello")

输出结果

__call()__ zhangsan
hello hello

2.ToTensor、Nomalize、Resize、Compose和RandomCrop等工具类的使用

六、torchvision中的数据集使用

七、DataLoader的使用

Dataset如果是一叠扑克牌的话,就是数据集,定义了其在哪里,量有多少等。DataLoader是一个加载器,就是一只手,参数就是告诉这只手怎么抓取扑克牌,如取几次,一次取多少张等。

DataLoader常用参数

  • dataset (Dataset) – 使用哪个数据集
  • batch_size (int, optional) – 一次抓取多少个数据,多少张牌
  • shuffle (bool, optional) – 是否重新洗牌,若true下次重新洗牌
  • num_workers (int, optional) – 多线程,windows下要设成0,不然会出错,出现

        BrokenPipeError: [Errno 32] Broken pipe

  • drop_last (bool, optional) – 当数据集大小不能被批大小整除时,设置为True则以删除最后一个不完整的批。False,则不删除,最后一批将变小。(默认值:False)

datasetloader加载批数据过程:

代码--dataloader.py

import torchvision

# 准备的测试数据集
from torch.utils.data import DataLoader
from torch.utils.tensorboard import SummaryWriter

test_data = torchvision.datasets.CIFAR10("./dataset", train=False, transform=torchvision.transforms.ToTensor())

test_loader = DataLoader(dataset=test_data, batch_size=64, shuffle=True, num_workers=0, drop_last=True)

# 测试数据集中第一张图片及target
img, target = test_data[0]
print(img.shape)
print(target)

writer = SummaryWriter("dataloader")
for epoch in range(2):
    step = 0
    for data in test_loader:
        imgs, targets = data
        # print(imgs.shape)
        # print(targets)
        writer.add_images("Epoch: {}".format(epoch), imgs, step)
        step = step + 1

writer.close()

 shuffle位true,代码设置抓了两次,第二次洗牌了。每小次抓了64张牌。

八、神经网络的基本骨架-torch.nn

torch.nn为神经网络提供骨架,我们往里面填充东西就是神经网络了。

1.Containers

代码-nn_module.py

import torch
from torch import nn


class Tudui(nn.Module):
    def __init__(self):
        super().__init__()

    def forward(self, input):
        output = input + 1
        return output


tudui = Tudui()
x = torch.tensor(1.0)
output = tudui(x)
print(output)

2.卷积操作

代码-nn_conv.py

import torch
import torch.nn.functional as F

input = torch.tensor([[1, 2, 0, 3, 1],
                      [0, 1, 2, 3, 1],
                      [1, 2, 1, 0, 0],
                      [5, 2, 3, 1, 1],
                      [2, 1, 0, 1, 1]])  #输入图像

kernel = torch.tensor([[1, 2, 1],
                       [0, 1, 0],
                       [2, 1, 0]])  #内核

input = torch.reshape(input, (1, 1, 5, 5))
kernel = torch.reshape(kernel, (1, 1, 3, 3))

print(input.shape)
print(kernel.shape)

output = F.conv2d(input, kernel, stride=1)  #输出图像
print(output)

output2 = F.conv2d(input, kernel, stride=2)
print(output2)

output3 = F.conv2d(input, kernel, stride=1, padding=1)
print(output3)

3.卷积层

代码-nn_conv2.py

dataset = torchvision.datasets.CIFAR10("../data", train=False, transform=torchvision.transforms.ToTensor(),
                                       download=True)
dataloader = DataLoader(dataset, batch_size=64)

class Tudui(nn.Module):
    def __init__(self):
        super(Tudui, self).__init__()
        self.conv1 = Conv2d(in_channels=3, out_channels=6, kernel_size=3, stride=1, padding=0)

    def forward(self, x):
        x = self.conv1(x)
        return x

tudui = Tudui()

writer = SummaryWriter("../logs")

step = 0
for data in dataloader:
    imgs, targets = data
    output = tudui(imgs)
    print(imgs.shape)
    print(output.shape)
    # torch.Size([64, 3, 32, 32])
    writer.add_images("input", imgs, step)
    # torch.Size([64, 6, 30, 30])  -> [xxx, 3, 30, 30]

    output = torch.reshape(output, (-1, 3, 30, 30))  #需要变成3个通道才能显示
    writer.add_images("output", output, step)

    step = step + 1

4.最大池化

作用:保留输入的主要特征,同时数据量减少,这样训练更快。代码---nn_maxpool.py


dataset = torchvision.datasets.CIFAR10("../data", train=False, download=True,
                                       transform=torchvision.transforms.ToTensor())

dataloader = DataLoader(dataset, batch_size=64)

class Tudui(nn.Module):
    def __init__(self):
        super(Tudui, self).__init__()
        self.maxpool1 = MaxPool2d(kernel_size=3, ceil_mode=False)  #ceil_mode=true

    def forward(self, input):
        output = self.maxpool1(input)
        return output

tudui = Tudui()

writer = SummaryWriter("../logs_maxpool")  
step = 0

for data in dataloader:
    imgs, targets = data
    writer.add_images("input", imgs, step)
    output = tudui(imgs)
    writer.add_images("output", output, step)
    step = step + 1

writer.close()

5.非线性激活

作用:非线性变换,可以给网络模型引入非线性特征,这样泛化更强。代码-nn_relu.py

dataset = torchvision.datasets.CIFAR10("../data", train=False, download=True,
                                       transform=torchvision.transforms.ToTensor())

dataloader = DataLoader(dataset, batch_size=64)

class Tudui(nn.Module):
    def __init__(self):
        super(Tudui, self).__init__()
        self.relu1 = ReLU()
        self.sigmoid1 = Sigmoid()

    def forward(self, input):
        output = self.sigmoid1(input)
        return output

tudui = Tudui()

writer = SummaryWriter("../logs_relu")
step = 0
for data in dataloader:
    imgs, targets = data
    writer.add_images("input", imgs, global_step=step)
    output = tudui(imgs)
    writer.add_images("output", output, step)
    step += 1

writer.close()

6.线层及其他层介绍

dataset = torchvision.datasets.CIFAR10("../data", train=False, transform=torchvision.transforms.ToTensor(),
                                      download=True)
dataloader = DataLoader(dataset, batch_size=64)
class Tudui(nn.Module):
    def __init__(self):
        super(Tudui, self).__init__()
        self.linear1 = Linear(196608, 10) # 输入特征个数,输出特征个数

    def forward(self,input):
        output =self.linear1(input)
        return  output
tudui =Tudui()
for data in dataloader:
    img, targets = data
    print(img.shape)
    output = torch.reshape(img, (1, 1, 1, -1))  #等价于 output=torch.flatten(img)
    print(output.shape)
    output= tudui(output)
    print(output.shape)

输出结果部分:

torch.Size([64, 3, 32, 32])
torch.Size([1, 1, 1, 196608])
torch.Size([1, 1, 1, 10])

7.搭建小实战和Sequential的使用

cifar10 model structure

根据框架中可以看出in_channels=3,out_channels=32,kernel_size=5 

但是由于经过5*5大小的卷积核之后,图片的大小依然是32*32,所以我们要计算padding和stride

conv2d的pytorch文件有公式可以套:

 代码--nn_seq.py

class Tudui(nn.Module):
    def __init__(self):
        super(Tudui, self).__init__()
        self.model1 = Sequential(
            Conv2d(3, 32, 5, padding=2),
            MaxPool2d(2),
            Conv2d(32, 32, 5, padding=2),
            MaxPool2d(2),
            Conv2d(32, 64, 5, padding=2),
            MaxPool2d(2),
            Flatten(),
            Linear(1024, 64),
            Linear(64, 10)
        )

    def forward(self, x):
        x = self.model1(x)
        return x

tudui = Tudui()
print(tudui)
input = torch.ones((64, 3, 32, 32))
output = tudui(input)
print(output.shape)

writer = SummaryWriter("../logs_seq")
writer.add_graph(tudui, input)
writer.close()

或者

class Tudui(nn.Module):
    def __init__(self):
        super(Tudui, self).__init__()
        self.conv1=Conv2d(3,32,5,padding=2)
        self.maxpool1 = MaxPool2d(2)
        self.conv2=Conv2d(32,32,5,padding=2)
        self.maxpool2 =MaxPool2d(2)
        self.conv3=Conv2d(32,64,5,padding=2)
        self.maxpool3 =MaxPool2d(2)
        self.flatten =Flatten()  #弄成一行1024列,64*4*4
        self.Linear1= Linear(1024,64)
        self.Linear2 = Linear(64, 10)

    def forward(self,x):
        x =self.conv1(x)
        x=self.maxpool1(x)
        x=self.conv2(x)
        x=self.maxpool2(x)
        x = self.conv3(x)
        x = self.maxpool3(x)
        x=self.flatten(x)
        x=self.Linear1(x)
        x=self.Linear2(x)
        return  x

#下列是打印网络结构
tudui=Tudui()
print(tudui)

#创建数据,检验网络结构
input = torch.ones((64,3,32,32))  #创建一个大小64,3个通道,32*32的图片
output =tudui(input)
print(output.shape)  #输出torch.Size([64, 10])为正确

#显示出网络结构
writer=SummaryWriter("logs")
writer.add_graph(tudui,input)
writer.close()

8.损失函数与反向传播

引入:好比一张试卷target是目标分数,output是实际分数

 Loss越小实际输出越接近目标值,所以Loss一般是越小越好。Loss调整参数,更新输出,使实际输出越接近目标值。

Loss Functions

Pytorch学习之十九种损失函数_mingo_敏的博客-CSDN博客

梯度下降 

cost函数是loss函数,梯度不断下降,不断更新参数,梯度为0,更新参数,此时产出的模型Loss是最小的。

反向传播

代码--nn_loss_network.py

dataset = torchvision.datasets.CIFAR10("../data", train=False, transform=torchvision.transforms.ToTensor(),
                                       download=True)

dataloader = DataLoader(dataset, batch_size=1)

class Tudui(nn.Module):
    def __init__(self):
        super(Tudui, self).__init__()
        self.model1 = Sequential(
            Conv2d(3, 32, 5, padding=2),
            MaxPool2d(2),
            Conv2d(32, 32, 5, padding=2),
            MaxPool2d(2),
            Conv2d(32, 64, 5, padding=2),
            MaxPool2d(2),
            Flatten(),
            Linear(1024, 64),
            Linear(64, 10)
        )

    def forward(self, x):
        x = self.model1(x)
        return x


loss = nn.CrossEntropyLoss()
tudui = Tudui()
for data in dataloader:
    imgs, targets = data
    outputs = tudui(imgs)
    result_loss = loss(outputs, targets)
    result_loss.backward() #反向传播
    print( result_loss)

9.优化器

损失函数,调用backward(),然后得到反向传播,就可以求出每个需要调节的参数,利用参数就可以求出梯度,有了梯度,就可以使用优化器对参数进行调整,以达到整体误差降低的目的。

代码--nn_optim.py

dataset = torchvision.datasets.CIFAR10("../data", train=False, transform=torchvision.transforms.ToTensor(),
                                       download=True)

dataloader = DataLoader(dataset, batch_size=1)

class Tudui(nn.Module):
    def __init__(self):
        super(Tudui, self).__init__()
        self.model1 = Sequential(
            Conv2d(3, 32, 5, padding=2),
            MaxPool2d(2),
            Conv2d(32, 32, 5, padding=2),
            MaxPool2d(2),
            Conv2d(32, 64, 5, padding=2),
            MaxPool2d(2),
            Flatten(),
            Linear(1024, 64),
            Linear(64, 10)
        )

    def forward(self, x):
        x = self.model1(x)
        return x


loss = nn.CrossEntropyLoss()
tudui = Tudui()
# 优化器通常设置模型的参数,学习率等
optim = torch.optim.SGD(tudui.parameters(), lr=0.01)
scheduler = StepLR(optim, step_size=5, gamma=0.1)
for epoch in range(20):   #训练多少轮
    running_loss = 0.0
    for data in dataloader:
        imgs, targets = data
        outputs = tudui(imgs)
        result_loss = loss(outputs, targets)
        optim.zero_grad()  #把上一次计算的梯度清零
        result_loss.backward()   # 损失反向传播,计算梯度
        scheduler.step()   # 使用梯度进行学习,即参数的优化
        running_loss = running_loss + result_loss  #训练多少轮的loss总和
    print(running_loss)

九、现有网络模型的使用及修改

代码--model_pretrained.py

vgg16_false = torchvision.models.vgg16(pretrained=False)#没有训练的模型,原始的神经网络参数
vgg16_true = torchvision.models.vgg16(pretrained=True)

# 增加网络结点
# vgg16_true.add_module('add_linear',nn.Linear(1000,10))
vgg16_true.classifier.add_module("add_linear", nn.Linear(1000, 10))
print(vgg16_true)

# 修改原有结构,不增加网络结点
vgg16_false.classifier[6] =nn.Linear(4096,10)
print(vgg16_false)

十、网络模型的保存与读取

代码见model_save.py和model_load.py

保存模型

import torch
import torchvision
from torch import nn

vgg16 = torchvision.models.vgg16(pretrained=False)
# 保存方式1,模型结构+模型参数
torch.save(vgg16, "vgg16_method1.pth")

# 保存方式2,模型参数(官方推荐)
torch.save(vgg16.state_dict(), "vgg16_method2.pth")

# 陷阱
class Tudui(nn.Module):
    def __init__(self):
        super(Tudui, self).__init__()
        self.conv1 = nn.Conv2d(3, 64, kernel_size=3)

    def forward(self, x):
        x = self.conv1(x)
        return x

tudui = Tudui()
torch.save(tudui, "tudui_method1.pth")

加载模型

import torch
from model_save import *
# 方式1-》保存方式1,加载模型
import torchvision
from torch import nn

model = torch.load("vgg16_method1.pth")
# print(model)

# 方式2,加载模型
vgg16 = torchvision.models.vgg16(pretrained=False)
vgg16.load_state_dict(torch.load("vgg16_method2.pth"))
# model = torch.load("vgg16_method2.pth")
# print(vgg16)

# 陷阱1
# class Tudui(nn.Module):
#     def __init__(self):
#         super(Tudui, self).__init__()
#         self.conv1 = nn.Conv2d(3, 64, kernel_size=3)
#
#     def forward(self, x):
#         x = self.conv1(x)
#         return x

model = torch.load('tudui_method1.pth')  #需要模型定义,陷阱1部分,或者从其他地方引入from model_save import *
print(model)

十一、完整的模型训练

网络结构--model.py

import torch
from torch import nn

# 搭建神经网络
class Tudui(nn.Module):
    def __init__(self):
        super(Tudui, self).__init__()
        self.model = nn.Sequential(
            nn.Conv2d(3, 32, 5, 1, 2),
            nn.MaxPool2d(2),
            nn.Conv2d(32, 32, 5, 1, 2),
            nn.MaxPool2d(2),
            nn.Conv2d(32, 64, 5, 1, 2),
            nn.MaxPool2d(2),
            nn.Flatten(),
            nn.Linear(64*4*4, 64),
            nn.Linear(64, 10)
        )

    def forward(self, x):
        x = self.model(x)
        return x


if __name__ == '__main__':
    tudui = Tudui()
    input = torch.ones((64, 3, 32, 32))
    output = tudui(input)
    print(output.shape)

训练代码--train.py

import torchvision
from torch.utils.tensorboard import SummaryWriter

from model import *
# 准备数据集
from torch import nn
from torch.utils.data import DataLoader
rain_data = torchvision.datasets.CIFAR10(root="../data", train=True, transform=torchvision.transforms.ToTensor(),
                                          download=True)
test_data = torchvision.datasets.CIFAR10(root="../data", train=False, transform=torchvision.transforms.ToTensor(),
                                         download=True)

# length 长度
train_data_size = len(train_data)
test_data_size = len(test_data)
# 如果train_data_size=10, 训练数据集的长度为:10
print("训练数据集的长度为:{}".format(train_data_size))
print("测试数据集的长度为:{}".format(test_data_size))


# 利用 DataLoader 来加载数据集
train_dataloader = DataLoader(train_data, batch_size=64)
test_dataloader = DataLoader(test_data, batch_size=64)

# 创建网络模型
tudui = Tudui()

# 损失函数
loss_fn = nn.CrossEntropyLoss()

# 优化器
# learning_rate = 0.01
# 1e-2=1 x (10)^(-2) = 1 /100 = 0.01
learning_rate = 1e-2
optimizer = torch.optim.SGD(tudui.parameters(), lr=learning_rate)

# 设置训练网络的一些参数
# 记录训练的次数
total_train_step = 0
# 记录测试的次数
total_test_step = 0
# 训练的轮数
epoch = 10

# 添加tensorboard
writer =SummaryWriter("logs")

for i in range(epoch):
    print("--------第{}轮训练开始--------".format(i))
    for data in train_dataloader:
        imgs, targets = data
        outputs = tudui(imgs)
        loss = loss_fn(outputs, targets)

        # 优化器优化模型
        optimizer.zero_grad()
        loss.backward()
        optimizer.step()

        total_train_step = total_train_step + 1
        if total_train_step % 100 == 0:
            print("训练次数:{}, Loss: {}".format(total_train_step, loss.item()))
            writer.add_scalar("train_loss", loss.item(), total_train_step)

    # 测试步骤开始
    total_test_loss=0
    with torch.no_grad():
        for data in test_dataloader:
            imgs,targets =data
            outputs = tudui(imgs)
            loss = loss_fn(outputs,targets)
            total_test_loss=total_test_loss+loss.item()
    print("整体测试集上的Loss:{}".format(total_test_loss))
    writer.add_scalar("test_loss", total_test_loss, total_test_step)
    total_test_step =total_test_step+1
    torch.save(tudui,"tudu_{}.pth".format(i))
    print("模型已保存")

writer.close()

但是测试集是不需要进行反向传播的,也就是说不需要计算去减小误差使模型拟合的,测试集的作用只是用来验证你模型效果的好坏,并不是用来训练模型用的,with torch.no_grad()就是防止进行梯度下降的。

准确率的计算

输出结果:

 

 或者训练代码--train.py

import torchvision
from torch.utils.tensorboard import SummaryWriter

from model import *
# 准备数据集
from torch import nn
from torch.utils.data import DataLoader
train_data = torchvision.datasets.CIFAR10(root="../data", train=True, transform=torchvision.transforms.ToTensor(),
                                          download=True)
test_data = torchvision.datasets.CIFAR10(root="../data", train=False, transform=torchvision.transforms.ToTensor(),
                                         download=True)

# length 长度
train_data_size = len(train_data)
test_data_size = len(test_data)
# 如果train_data_size=10, 训练数据集的长度为:10
print("训练数据集的长度为:{}".format(train_data_size))
print("测试数据集的长度为:{}".format(test_data_size))


# 利用 DataLoader 来加载数据集
train_dataloader = DataLoader(train_data, batch_size=64)
test_dataloader = DataLoader(test_data, batch_size=64)

# 创建网络模型
tudui = Tudui()

# 损失函数
loss_fn = nn.CrossEntropyLoss()

# 优化器
# learning_rate = 0.01
# 1e-2=1 x (10)^(-2) = 1 /100 = 0.01
learning_rate = 1e-2
optimizer = torch.optim.SGD(tudui.parameters(), lr=learning_rate)

# 设置训练网络的一些参数
# 记录训练的次数
total_train_step = 0
# 记录测试的次数
total_test_step = 0
# 训练的轮数
epoch = 10

# 添加tensorboard
writer = SummaryWriter("../logs_train")

for i in range(epoch):
    print("-------第 {} 轮训练开始-------".format(i+1))

    # 训练步骤开始
    tudui.train()
    for data in train_dataloader:
        imgs, targets = data
        outputs = tudui(imgs)
        loss = loss_fn(outputs, targets)

        # 优化器优化模型
        optimizer.zero_grad()
        loss.backward()
        optimizer.step()

        total_train_step = total_train_step + 1
        if total_train_step % 100 == 0:
            print("训练次数:{}, Loss: {}".format(total_train_step, loss.item()))
            writer.add_scalar("train_loss", loss.item(), total_train_step)

    # 测试步骤开始
    tudui.eval()
    total_test_loss = 0
    total_accuracy = 0
    with torch.no_grad():
        for data in test_dataloader:
            imgs, targets = data
            outputs = tudui(imgs)
            loss = loss_fn(outputs, targets)
            total_test_loss = total_test_loss + loss.item()
            accuracy = (outputs.argmax(1) == targets).sum()
            total_accuracy = total_accuracy + accuracy

    print("整体测试集上的Loss: {}".format(total_test_loss))
    print("整体测试集上的正确率: {}".format(total_accuracy/test_data_size))
   
    writer.add_scalar("test_accuracy", total_accuracy/test_data_size, total_test_step)
    total_test_step = total_test_step + 1

    torch.save(tudui, "tudui_{}.pth".format(i))
    print("模型已保存")

writer.close()

十二、利用GPU训练

1.方法一

有三种 

可以使用谷歌GPU

2.方法二

#方法一

tudui = Tudui()
if torch.cuda.is_available():
    tudui = tudui.cuda()


#方法二
# 定义训练的设备
device = torch.device("cuda")
#使用
tudui = Tudui()
tudui = tudui.to(device)

 

 十三、模型验证套路

import torch
import torchvision
from PIL import Image
from torch import nn

image_path = "../imgs/airplane.png"
image = Image.open(image_path)
print(image)
image = image.convert('RGB')
transform = torchvision.transforms.Compose([torchvision.transforms.Resize((32, 32)),
                                            torchvision.transforms.ToTensor()])

image = transform(image)
print(image.shape)

class Tudui(nn.Module):
    def __init__(self):
        super(Tudui, self).__init__()
        self.model = nn.Sequential(
            nn.Conv2d(3, 32, 5, 1, 2),
            nn.MaxPool2d(2),
            nn.Conv2d(32, 32, 5, 1, 2),
            nn.MaxPool2d(2),
            nn.Conv2d(32, 64, 5, 1, 2),
            nn.MaxPool2d(2),
            nn.Flatten(),
            nn.Linear(64*4*4, 64),
            nn.Linear(64, 10)
        )

    def forward(self, x):
        x = self.model(x)
        return x

model = torch.load("tudui_29_gpu.pth", map_location=torch.device('cpu'))
print(model)
image = torch.reshape(image, (1, 3, 32, 32))  #一张图片
model.eval()
with torch.no_grad():
    output = model(image)
print(output)

print(output.argmax(1))

  • 0
    点赞
  • 0
    收藏
    觉得还不错? 一键收藏
  • 0
    评论

“相关推荐”对你有帮助么?

  • 非常没帮助
  • 没帮助
  • 一般
  • 有帮助
  • 非常有帮助
提交
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值