用两个优化器独立训练神经网络两个模块

创建了一个ResNet模型并设置了两个不同的损失函数和优化器,分别用于训练卷积层和全连接层。在训练循环中,对输入数据进行前向传播,计算损失,然后反向传播更新权重。使用了Adam优化器,学习率为0.001。
摘要由CSDN通过智能技术生成
# 创建 ResNet 实例
resnet = ResNet(num_classes, args)

# 定义损失函数和优化器
loss_fn1 = nn.CrossEntropyLoss()
loss_fn2 = nn.MSELoss()
optimizer1 = optim.Adam([
    {'params': resnet.conv1.parameters()},
    {'params': resnet.conv2.parameters()},
    {'params': resnet.cross_attention.parameters()}
], lr=0.001)                        #调整自己想要的学习率
optimizer2 = optim.Adam([
    {'params': resnet.backbone.parameters()},
    {'params': resnet.fc.parameters()},
    {'params': resnet.fc2.parameters()}
], lr=0.001)

# 训练循环
for epoch in range(num_epochs):
    total_loss1 = 0.0
    total_loss2 = 0.0

    for inputs, targets in dataloader:
        optimizer1.zero_grad()
        optimizer2.zero_grad()

        outputs1, outputs2 = resnet(inputs)
        loss1 = loss_fn1(outputs1, targets)
        loss2 = loss_fn2(outputs2, targets)

        loss1.backward()
        loss2.backward()

        optimizer1.step()
        optimizer2.step()

  • 2
    点赞
  • 4
    收藏
    觉得还不错? 一键收藏
  • 0
    评论
以下是基于MindSpore的带有两个残差模块的AlexNet神经网络的代码: ```python import mindspore.nn as nn import mindspore.ops.operations as P import mindspore.common.dtype as mstype class AlexNet(nn.Cell): def __init__(self, num_classes=10): super(AlexNet, self).__init__() self.conv1 = nn.Conv2d(in_channels=3, out_channels=64, kernel_size=11, stride=4, pad_mode='valid') self.relu1 = nn.ReLU() self.pool1 = nn.MaxPool2d(kernel_size=3, stride=2, pad_mode='valid') self.conv2 = nn.Conv2d(in_channels=64, out_channels=192, kernel_size=5, pad_mode='same') self.relu2 = nn.ReLU() self.pool2 = nn.MaxPool2d(kernel_size=3, stride=2, pad_mode='valid') self.conv3 = nn.Conv2d(in_channels=192, out_channels=384, kernel_size=3, pad_mode='same') self.relu3 = nn.ReLU() self.conv4 = nn.Conv2d(in_channels=384, out_channels=256, kernel_size=3, pad_mode='same') self.relu4 = nn.ReLU() self.conv5 = nn.Conv2d(in_channels=256, out_channels=256, kernel_size=3, pad_mode='same') self.relu5 = nn.ReLU() self.pool5 = nn.MaxPool2d(kernel_size=3, stride=2, pad_mode='valid') self.flatten = nn.Flatten() self.fc1 = nn.Dense(in_channels=6*6*256, out_channels=4096) self.relu6 = nn.ReLU() self.fc2 = nn.Dense(in_channels=4096, out_channels=4096) self.relu7 = nn.ReLU() self.fc3 = nn.Dense(in_channels=4096, out_channels=num_classes) self.softmax = nn.Softmax(axis=1) self.add = P.TensorAdd() def construct(self, x): x = self.pool1(self.relu1(self.conv1(x))) x = self.pool2(self.relu2(self.conv2(x))) x = self.relu3(self.conv3(x)) x = self.relu4(self.conv4(x)) x = self.pool5(self.relu5(self.conv5(x))) x = self.flatten(x) x = self.relu6(self.fc1(x)) x = self.relu7(self.fc2(x)) x = self.fc3(x) x1 = x x = self.fc3(self.relu7(self.fc2(self.relu6(self.fc1(x))))) x = self.add(x, x1) return self.softmax(x) ``` 其中,我们在fc1和fc2之间添加了一个残差模块,在fc2之后再添加一个残差模块。这两个残差模块的实现方式都是使用TensorAdd来实现的。
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值