PyTorch项目应用实例(二)ResNet | SENet实现coco多标签分类

背景:之前的网络结构做多标签分类较为复杂,我们需要用resNet进行多标签分类简单的实现相应的分类问题,测试baseline。

目录

一、网络结构及定义

二、optimizer

三、新定义网络

3.1 加载模型

3.2 模型定义

3.3 网络尺寸

3.4 fc层尺寸

四、训练内存占用

五、SENet更改

5.1 引入模型

5.2 SENet定义代码


一、网络结构及定义

在原来代码基础上更改,方便运行。

加入resnet101的结构

    elif Config.MODEL == 'resnet101':
        model = models.resnet101(pretrained=False)
        print('load pretrained model...')
        model.load_state_dict(torch.load('./resnet101-5d3b4d8f.pth'))
    # model params
    MODEL = 'resnet101'  # options: hgat_conv, hgat_fc, groupnet
    BACKBONE = 'resnet101'
    GROUPS = 12

二、optimizer

报错:

    optimizer = torch.optim.SGD(model.parameters(), 'lr':parser.lr,
                                                        ^
SyntaxError: invalid syntax

之前的optimizer

optimizer = torch.optim.SGD(model.get_config_optim(args.lr, args.lrp),
                                    # lr=args.lr,
                                    momentum=args.momentum,
                                    weight_decay=args.weight_decay)


    def get_config_optim(self, lr, lrp):
        return [
            {'params': self.features.parameters(), 'lr': lrp},
            {'params': self.heads.parameters(), 'lr': lr},
        ]

加入resnet之后,相应的optimizer不能用之前的了。因为没有了相应的get_config_optim

    if Config.MODEL == 'resnet101':
        optimizer = torch.optim.SGD(params=model.parameters(), lr=args.lr,
                                    # lr=args.lr,
                                    momentum=args.momentum,
                                    weight_decay=args.weight_decay)
    else:
        optimizer = torch.optim.SGD(model.get_config_optim(args.lr, args.lrp),
                                    # lr=args.lr,
                                    momentum=args.momentum,
                                    weight_decay=args.weight_decay)

绕进去之后,发现一层一层调试下去,最终的resnet输入与之前的不一样

  File "/home/xingxiangrui/chun-ML_GCN/engine.py", line 256, in train
    self.on_forward(True, model, criterion, data_loader, optimizer)
  File "/home/xingxiangrui/chun-ML_GCN/engine.py", line 446, in on_forward
    self.state['output'] = model(feature_var, inp_var)
  File "/home/xingxiangrui/chun-ML_GCN/env/lib/python3.6/site-packages/torch/nn/modules/module.py", line 357, in __call__
    result = self.forward(*input, **kwargs)
  File "/home/xingxiangrui/chun-ML_GCN/env/lib/python3.6/site-packages/torch/nn/parallel/data_parallel.py", line 73, in forward
    outputs = self.parallel_apply(replicas, inputs, kwargs)
。。。
  File "/home/xingxiangrui/chun-ML_GCN/env/lib/python3.6/site-packages/torch/nn/modules/module.py", line 357, in __call__
    result = self.forward(*input, **kwargs)
TypeError: forward() takes 2 positional arguments but 3 were given

所以,改为采用新网络,重新定义ml_resnet.py函数,然后import进来。

三、新定义网络

3.1 加载模型

    elif Config.MODEL == 'resnet101':
        import mymodels.ml_resnet as ml_resnet
        model= ml_resnet.ML_RESNET(Config.BACKBONE, groups=Config.GROUPS, nclasses=Config.NCLASSES,
                            nclasses_per_group=Config.NCLASSES_PER_GROUP,
                            group_channels=Config.GROUP_CHANNELS, class_channels=Config.CLASS_CHANNELS)

 

3.2 模型定义

从文件中加载定义,创建新文件

class ML_RESNET(nn.Module):
    def __init__(self, backbone, groups, nclasses, nclasses_per_group, group_channels, class_channels):
        super(HGAT_FC, self).__init__()
        self.groups = groups
        self.nclasses = nclasses
        self.nclasses_per_group = nclasses_per_group
        self.group_channels = group_channels
        self.class_channels = class_channels
        if backbone == 'resnet101':
            model = models.resnet101(pretrained=False)
            print('load pretrained model...')
            model.load_state_dict(torch.load('./resnet101-5d3b4d8f.pth'))
        elif backbone == 'resnet50':
            model = models.resnet50(pretrained=False)
            print('load pretrained model...')
            model.load_state_dict(torch.load('./resnet50-5d3b4d8f.pth'))
        else:
            raise Exception()
        self.features = nn.Sequential(
            model.conv1,
            model.bn1,
            model.relu,
            model.maxpool,
            model.layer1,
            model.layer2,
            model.layer3,
            model.layer4, )
        self.gmp = nn.AdaptiveMaxPool2d(1)
        self.fc=nn.Sequential(utils.BasicLinear(in_channels=2048, out_channels=1024),
                                       utils.BasicLinear(in_channels=1024, out_channels=class_channels), )

        self.image_normalization_mean = [0.485, 0.456, 0.406]
        self.image_normalization_std = [0.229, 0.224, 0.225]

    def forward(self, x, inp):
        x = self.features(x)  # [B,2048,H,W]
        x=self.gmp(x).view(x.size(0),x.size(1))
        x=self.fc(x)
        return x

    def get_config_optim(self, lr, lrp):
        return [
            {'params': self.features.parameters(), 'lr': lrp},
            #{'params': self.heads.parameters(), 'lr': lr},
        ]

定义需要改成自己的定义,ML_RESNET

class ML_RESNET(nn.Module):
    def __init__(self, backbone, groups, nclasses, nclasses_per_group, group_channels, class_channels):
        super(ML_RESNET, self).__init__()

3.3 网络尺寸

之前的网络尺寸定义为:

            raise Exception()
        self.features = nn.Sequential(
            model.conv1,
            model.bn1,
            model.relu,
            model.maxpool,
            model.layer1,
            model.layer2,
            model.layer3,
            model.layer4, )
        self.gmp = nn.AdaptiveMaxPool2d(1)
        self.fc=nn.Sequential(utils.BasicLinear(in_channels=2048, out_channels=1024),
                                       utils.BasicLinear(in_channels=1024, out_channels=class_channels), )
    def forward(self, x, inp):
        x = self.features(x)  # [B,2048,H,W]
        x=self.gmp(x).view(x.size(0),x.size(1))
        x=self.fc(x)
        return x

结果产生报错

  File "general_train.py", line 182, in <module>
    main_coco()
  File "general_train.py", line 178, in main_coco
    engine.learning(model, criterion, train_dataset, val_dataset, optimizer)
  File "/home/xingxiangrui/chun-ML_GCN/engine.py", line 214, in learning
    self.train(train_loader, model, criterion, optimizer, epoch)
  File "/home/xingxiangrui/chun-ML_GCN/engine.py", line 256, in train
    self.on_forward(True, model, criterion, data_loader, optimizer)
  File "/home/xingxiangrui/chun-ML_GCN/engine.py", line 451, in on_forward
    weight=torch.autograd.Variable(weights.cuda()))
  File "/home/xingxiangrui/chun-ML_GCN/env/lib/python3.6/site-packages/torch/nn/functional.py", line 1227, in binary_cross_entropy_with_logits
    raise ValueError("Target size ({}) must be the same as input size ({})".format(target.size(), input.size()))
ValueError: Target size (torch.Size([32, 80])) must be the same as input size (torch.Size([32, 256]))

出现这个错误的原因,是这样,,一个为80,一个为256

看报错,是交叉熵的过程中出现的错误

    NCLASSES = 80
    NCLASSES_PER_GROUP = [1, 8, 5, 10, 5, 10, 7, 10, 6, 6, 5, 7]  # FIXME: to check
    GROUP_CHANNELS = 512
    CLASS_CHANNELS = 256

class channels与class numbers出现了错误,其中一个为256一个为80

定义时进行更改

        self.features = nn.Sequential(
            model.conv1,
            model.bn1,
            model.relu,
            model.maxpool,
            model.layer1,
            model.layer2,
            model.layer3,
            model.layer4, )
        self.gmp = nn.AdaptiveMaxPool2d(1)
        self.fc=nn.Sequential(utils.BasicLinear(in_channels=2048, out_channels=1024),
                                       utils.BasicLinear(in_channels=1024, out_channels=nclasses), )

将最终输出结果改为nclasses

3.4 fc层尺寸

之前fc层选用错了,不该用basicLinear

        self.fc=nn.Sequential(utils.BasicLinear(in_channels=2048, out_channels=1024),
                                       utils.BasicLinear(in_channels=1024, out_channels=nclasses), )

basiclinear是我们自己定义的。

class BasicLinear(nn.Module):
    def __init__(self, in_channels, out_channels):
        super(BasicLinear, self).__init__()
        self.fc = nn.Linear(in_features=in_channels, out_features=out_channels, bias=False)
        self.bn = nn.BatchNorm1d(num_features=out_channels)
        self.relu = nn.ReLU()

加了ReLU可能导致无法收敛,我们应当直接加linear层来保证模型收敛。

更改self.fc尺寸

        # self.fc=nn.Sequential(utils.BasicLinear(in_channels=2048, out_channels=1024),
        #                                utils.BasicLinear(in_channels=1024, out_channels=nclasses), )
        self.fc = nn.Linear(in_features=2048, out_features=nclasses, bias=True)

四、训练内存占用

训练过程依然占据四张显卡。且占用内存基本不变。

若用三张显卡训练,会报错,list out of memroy

[xingxiangrui@gzbh-mms-gpu55.gzbh.baidu.com chun-ML_GCN]$ nvidia-smi
Sun May  5 16:38:47 2019
+-----------------------------------------------------------------------------+
| NVIDIA-SMI 384.81                 Driver Version: 384.81                    |
|-------------------------------+----------------------+----------------------+
| GPU  Name        Persistence-M| Bus-Id        Disp.A | Volatile Uncorr. ECC |
| Fan  Temp  Perf  Pwr:Usage/Cap|         Memory-Usage | GPU-Util  Compute M. |
|===============================+======================+======================|
|   0  Tesla P4            On   | 00000000:02:00.0 Off |                    0 |
| N/A   61C    P0    64W /  75W |   5711MiB /  7606MiB |     47%      Default |
+-------------------------------+----------------------+----------------------+
|   1  Tesla P4            On   | 00000000:03:00.0 Off |                    0 |
| N/A   62C    P0    62W /  75W |   5349MiB /  7606MiB |     71%      Default |
+-------------------------------+----------------------+----------------------+
|   2  Tesla P4            On   | 00000000:82:00.0 Off |                    0 |
| N/A   58C    P0    65W /  75W |   5375MiB /  7606MiB |     79%      Default |
+-------------------------------+----------------------+----------------------+
|   3  Tesla P4            On   | 00000000:83:00.0 Off |                    0 |
| N/A   59C    P0    63W /  75W |   5351MiB /  7606MiB |     38%      Default |
+-------------------------------+----------------------+----------------------+

可以用watch -n 1 nvidia-smi指令,查看显卡动态占用。

五、SENet更改

背景:需要将模型之中加入SENet的内容,因此需要更改骨架。

SENet的PyTorch代码,此代码没有预训练:https://github.com/moskomule/senet.pytorch

加入新的有预训练的代码:

5.1 引入模型

从模型引入相应的class

import models.senet_origin as senet_origin

文件代码地址见:SENet的PyTorch代码:https://github.com/moskomule/senet.pytorch

带有预训练的SENet的PyTorch代码:https://github.com/Xingxiangrui/various_pyTorch_network_structure/blob/master/senet_and_pretrained.py

        # fixme new SE-resnet backbone
        if backbone == 'resnet101':
            model = senet_origin.se_resnet101()
        elif backbone == 'resnet50':
            model = senet_origin.se_resnet50()
        elif backbone == 'resnet101_cbam':
            import mymodels.cbam as cbam
            model = cbam.resnet101_cbam()
        elif backbone=='resnet150':
            model=senet_origin.se_resnet152()
        else:
            raise Exception()
        # self.features = nn.Sequential(
        #     model.conv1,
        #     model.bn1,
        #     model.relu,
        #     model.maxpool,
        #     model.layer1,
        #     model.layer2,
        #     model.layer3,
        #     model.layer4, )
        self.features = nn.Sequential(
            # model.layer0,
            model.layer0,
            model.layer1,
            model.layer2,
            model.layer3,
            model.layer4 )

5.2 SENet定义代码

https://github.com/Xingxiangrui/various_pyTorch_network_structure/blob/master/senet_and_pretrained.py

直接从这里import即可。

  • 8
    点赞
  • 34
    收藏
    觉得还不错? 一键收藏
  • 打赏
    打赏
  • 5
    评论
这里是一个使用PyTorch实现ResNet18和SENet的代码,用于故障特征分类: ```python import torch import torch.nn as nn import torch.nn.functional as F class SEBlock(nn.Module): def __init__(self, channels, reduction=16): super(SEBlock, self).__init__() self.avg_pool = nn.AdaptiveAvgPool2d(1) self.fc1 = nn.Linear(channels, channels // reduction) self.fc2 = nn.Linear(channels // reduction, channels) def forward(self, x): b, c, _, _ = x.size() y = self.avg_pool(x).view(b, c) y = self.fc1(y) y = F.relu(y) y = self.fc2(y) y = F.sigmoid(y).view(b, c, 1, 1) return x * y.expand_as(x) class ResNet18(nn.Module): def __init__(self, in_channels=16, num_classes=2): super(ResNet18, self).__init__() self.conv1 = nn.Conv2d(in_channels, 64, kernel_size=3, stride=1, padding=1, bias=False) self.bn1 = nn.BatchNorm2d(64) self.relu = nn.ReLU(inplace=True) self.maxpool = nn.MaxPool2d(kernel_size=3, stride=2, padding=1) self.layer1 = self._make_layer(64, 2) self.layer2 = self._make_layer(128, 2, stride=2) self.layer3 = self._make_layer(256, 2, stride=2) self.layer4 = self._make_layer(512, 2, stride=2) self.avgpool = nn.AdaptiveAvgPool2d((1, 1)) self.fc = nn.Linear(512, num_classes) def _make_layer(self, planes, blocks, stride=1): downsample = None if stride != 1 or planes != 64: downsample = nn.Sequential( nn.Conv2d(64, planes, kernel_size=1, stride=stride, bias=False), nn.BatchNorm2d(planes) ) layers = [] layers.append(nn.Conv2d(64, planes, kernel_size=3, stride=stride, padding=1, bias=False)) layers.append(nn.BatchNorm2d(planes)) layers.append(nn.ReLU(inplace=True)) layers.append(nn.Conv2d(planes, planes, kernel_size=3, stride=1, padding=1, bias=False)) layers.append(nn.BatchNorm2d(planes)) layers.append(SEBlock(planes)) layers.append(nn.ReLU(inplace=True)) return nn.Sequential(*layers) def forward(self, x): x = self.conv1(x) x = self.bn1(x) x = self.relu(x) x = self.maxpool(x) x = self.layer1(x) x = self.layer2(x) x = self.layer3(x) x = self.layer4(x) x = self.avgpool(x) x = x.view(x.size(0), -1) x = self.fc(x) return x ``` 使用时,可以像这样创建一个ResNet18和SENet的结合模型: ```python model = ResNet18(in_channels=16, num_classes=10) ``` 其中,`in_channels=16`表示输入图片的通道数为16,`num_classes=10`表示分类的类别数为10。可以根据需要进行修改。

“相关推荐”对你有帮助么?

  • 非常没帮助
  • 没帮助
  • 一般
  • 有帮助
  • 非常有帮助
提交
评论 5
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包

打赏作者

祥瑞Coding

你的鼓励将是我创作的最大动力

¥1 ¥2 ¥4 ¥6 ¥10 ¥20
扫码支付:¥1
获取中
扫码支付

您的余额不足,请更换扫码支付或充值

打赏作者

实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值