paddle实现手写数字模型(一)

本文介绍了如何使用PaddlePaddle库在Python环境下构建LeNet神经网络模型,对MNIST数据集进行训练,并演示了如何加载预训练模型对本地图像进行预测。
摘要由CSDN通过智能技术生成
  1. 参考文档:paddle官网文档
  2. 环境:Python 3.12.2 ,pip 24.0 ,paddlepaddle 2.6.0
    python -m pip install paddlepaddle==2.6.0 -i https://pypi.tuna.tsinghua.edu.cn/simple
  3. 调试代码如下:
    LeNet.py
import paddle
import paddle.nn.functional as F

class LeNet(paddle.nn.Layer):
    def __init__(self):
        super().__init__()
        self.conv1 = paddle.nn.Conv2D(in_channels=1,out_channels=6,kernel_size=5,stride=1,padding=2)
        self.max_pool1 = paddle.nn.MaxPool2D(kernel_size=2,  stride=2)
        self.conv2 = paddle.nn.Conv2D(in_channels=6, out_channels=16, kernel_size=5, stride=1)
        self.max_pool2 = paddle.nn.MaxPool2D(kernel_size=2, stride=2)
        self.linear1 = paddle.nn.Linear(in_features=16*5*5, out_features=120)
        self.linear2 = paddle.nn.Linear(in_features=120, out_features=84)
        self.linear3 = paddle.nn.Linear(in_features=84, out_features=10)

    def forward(self, x):
        x = self.conv1(x)
        x = F.relu(x)
        x = self.max_pool1(x)
        x = self.conv2(x)
        x = F.relu(x)
        x = self.max_pool2(x)
        x = paddle.flatten(x, start_axis=1,stop_axis=-1)
        x = self.linear1(x)
        x = F.relu(x)
        x = self.linear2(x)
        x = F.relu(x)
        x = self.linear3(x)
        return x

train.py


import paddle
from paddle.vision.transforms import Compose,Normalize,ToTensor
import paddle.vision.transforms as T  

import numpy as np
import matplotlib.pyplot as plt
from paddle.metric import Accuracy

from LeNet import LeNet
from PIL import Image



print(paddle.__version__)
transform = Compose([Normalize(mean=[127.5],std=[127.5],data_format='CHW')])
print('下载和加载训练数据...')
train_dataset = paddle.vision.datasets.MNIST(mode='train',transform=transform)
test_dataset = paddle.vision.datasets.MNIST(mode='test',transform=transform)
print('load finished')

train_data0,train_label_0 = train_dataset[0][0],train_dataset[0][1]
train_data0 = train_data0.reshape([28,28])
plt.figure(figsize=(2,2))
plt.imshow(train_data0,cmap=plt.cm.binary)
#plt.show()
print('train_data0 label is: '+str(train_label_0))


model = paddle.Model(LeNet())   # 用Model封装模型
optim = paddle.optimizer.Adam(learning_rate=0.001, parameters=model.parameters())

# 配置模型
print('配置模型...')
model.prepare(
    optim,
    paddle.nn.CrossEntropyLoss(),
    Accuracy()
    )
# 训练模型
print('训练模型...')
model.fit(train_dataset,
        epochs=2,
        batch_size=64,
        verbose=1
        )
# 保存模型  
model.save('./model/mnist_model')  # 默认保存模型结构和参数 

#预测模型
print('预测模型...')
model.evaluate(test_dataset, batch_size=64, verbose=1)


predicted.py


import paddle

import numpy as np

from LeNet import LeNet
from PIL import Image

# 读取一张本地的样例图片,转变成模型输入的格式
def load_image(img_path):
    # 从img_path中读取图像,并转为灰度图
    im = Image.open(img_path).convert('L')
    #plt.imshow(im,cmap='gray')
    # print(np.array(im))
    im = im.resize((28, 28), Image.Resampling.LANCZOS)
    im = np.array(im).reshape(1, 1, 28, 28).astype(np.float32)
    # 图像归一化,保持和数据集的数据范围一致
    im = 1 - im / 255 
    return im

# 加载训练好的模型参数
model = LeNet()
model.load_dict(paddle.load('./model/mnist_model.pdparams'))

# 设置模型为评估模式
model.eval()

# 准备一个MNIST样例图像
example_image = load_image("d:/8.png")

# 转换为Tensor并进行推理
with paddle.no_grad():
    example_tensor = paddle.to_tensor(example_image)
    prediction = model(example_tensor)
    print(prediction)

# 获取预测类别
predicted_class = np.argmax(prediction.numpy(), axis=1)[0]
print(f"Predicted class: {predicted_class}")

说明:先通过执行train.py训练数据集,将模型保存在model文件夹中,
然后运行predicted.py加载训练出来的数据集,推理出d:/8.png图片的结果。
结果图片如下:
在这里插入图片描述

  • 3
    点赞
  • 3
    收藏
    觉得还不错? 一键收藏
  • 打赏
    打赏
  • 0
    评论
好的,以下是一个利用 VGG 模型解决手写数字识别的 Paddle 代码示例: ```python import paddle import paddle.nn as nn class VGGBlock(nn.Layer): def __init__(self, in_channels, out_channels, num_convs): super(VGGBlock, self).__init__() self.conv1 = nn.Conv2D(in_channels, out_channels, kernel_size=3, padding=1) self.relu = nn.ReLU() self.convs = nn.Sequential() for i in range(num_convs - 1): conv_name = 'conv{}'.format(i + 2) setattr(self.convs, conv_name, nn.Conv2D(out_channels, out_channels, kernel_size=3, padding=1)) self.pool = nn.MaxPool2D(kernel_size=2, stride=2) def forward(self, x): x = self.conv1(x) x = self.relu(x) x = self.convs(x) x = self.pool(x) return x class VGGNet(nn.Layer): def __init__(self): super(VGGNet, self).__init__() self.block1 = VGGBlock(in_channels=1, out_channels=64, num_convs=2) self.block2 = VGGBlock(in_channels=64, out_channels=128, num_convs=2) self.block3 = VGGBlock(in_channels=128, out_channels=256, num_convs=3) self.block4 = VGGBlock(in_channels=256, out_channels=512, num_convs=3) self.block5 = VGGBlock(in_channels=512, out_channels=512, num_convs=3) self.flatten = nn.Flatten() self.fc1 = nn.Linear(in_features=512*7*7, out_features=4096) self.fc2 = nn.Linear(in_features=4096, out_features=4096) self.fc3 = nn.Linear(in_features=4096, out_features=10) def forward(self, x): x = self.block1(x) x = self.block2(x) x = self.block3(x) x = self.block4(x) x = self.block5(x) x = self.flatten(x) x = self.fc1(x) x = self.fc2(x) x = self.fc3(x) return x # 加载数据集 train_dataset = paddle.vision.datasets.MNIST(mode='train') test_dataset = paddle.vision.datasets.MNIST(mode='test') # 定义数据预处理函数 transform = paddle.vision.transforms.Compose([ paddle.vision.transforms.Resize((32, 32)), paddle.vision.transforms.ToTensor(), paddle.vision.transforms.Normalize([0.5], [0.5]) ]) # 定义数据加载器 train_loader = paddle.io.DataLoader(train_dataset.transform(transform), batch_size=64, shuffle=True) test_loader = paddle.io.DataLoader(test_dataset.transform(transform), batch_size=64, shuffle=False) # 定义模型、优化器和损失函数 model = VGGNet() optimizer = paddle.optimizer.Adam(learning_rate=0.001, parameters=model.parameters()) criterion = nn.CrossEntropyLoss() # 训练模型 epochs = 10 for epoch in range(epochs): for batch_id, data in enumerate(train_loader()): x_data = data[0] y_data = paddle.to_tensor(data[1]) logits = model(x_data) loss = criterion(logits, y_data) loss.backward() optimizer.step() optimizer.clear_grad() if batch_id % 100 == 0: print("epoch: {}, batch_id: {}, loss is: {}".format(epoch, batch_id, loss.numpy())) # 评估模型 model.eval() accs = [] for batch_id, data in enumerate(test_loader()): x_data = data[0] y_data = paddle.to_tensor(data[1]) logits = model(x_data) pred = paddle.argmax(logits, axis=-1).numpy() acc = np.mean((pred == y_data.numpy()).astype(np.float32)) accs.append(acc) print("accuracy: ", np.mean(accs)) ``` 这份代码使用了 VGGNet 作为模型,并使用了 MNIST 数据集进行训练和测试。在训练过程中,采用了 Adam 优化器和交叉熵损失函数,训练了 10 个 epoch,最终在测试集上达到了较好的准确率。
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包

打赏作者

奔走的小龙虾

你的鼓励将是我创作的最大动力

¥1 ¥2 ¥4 ¥6 ¥10 ¥20
扫码支付:¥1
获取中
扫码支付

您的余额不足,请更换扫码支付或充值

打赏作者

实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值