下面是一个使用PyTorch库实现的ResNet-18的示例代码。为了简化代码,这里使用了PyTorch官方提供的torchvision.models
模块中的resnet18
预定义模型。但是,我也会为你提供一个简化的ResNet块和ResNet-18模型结构的示例代码,以便你了解ResNet的内部结构。
首先,使用预定义的ResNet-18模型:
python复制代码
import torch | |
import torchvision.models as models | |
# 使用预定义的ResNet-18模型 | |
resnet18 = models.resnet18(pretrained=False) # 设置为True将加载预训练权重 | |
# 打印模型结构 | |
print(resnet18) | |
# 假设你有一个输入张量x | |
# x = torch.randn(1, 3, 224, 224) # (batch_size, channels, height, width) | |
# output = resnet18(x) | |
# print(output.shape) # 应该会输出类似 torch.Size([1, 1000]) 的形状,表示1000个类别的输出 |
现在,如果你想要从头开始实现ResNet-18的一个简化版本(仅用于说明ResNet块和结构),以下是一个示例:
python复制代码
import torch | |
import torch.nn as nn | |
import torch.nn.functional as F | |
class BasicBlock(nn.Module): | |
expansion = 1 | |
def __init__(self, in_planes, planes, stride=1): | |
super(BasicBlock, self).__init__() | |
self.conv1 = nn.Conv2d( | |
in_planes, planes, kernel_size=3, stride=stride, padding=1, bias=False) | |
self.bn1 = nn.BatchNorm2d(planes) | |
self.conv2 = nn.Conv2d(planes, planes, kernel_size=3, stride=1, padding=1, bias=False) | |
self.bn2 = nn.BatchNorm2d(planes) | |
self.shortcut = nn.Sequential() | |
if stride != 1 or in_planes != self.expansion * planes: | |
self.shortcut = nn.Sequential( | |
nn.Conv2d(in_planes, self.expansion * planes, kernel_size=1, stride=stride, bias=False), | |
nn.BatchNorm2d(self.expansion * planes) | |
) | |
def forward(self, x): | |
out = F.relu(self.bn1(self.conv1(x))) | |
out = self.bn2(self.conv2(out)) | |
out += self.shortcut(x) | |
out = F.relu(out) | |
return out | |
class ResNet18(nn.Module): | |
def __init__(self, num_classes=1000): | |
super(ResNet18, self).__init__() | |
self.in_planes = 64 | |
self.conv1 = nn.Conv2d(3, 64, kernel_size=7, stride=2, padding=3, bias=False) | |
self.bn1 = nn.BatchNorm2d(64) | |
self.relu = nn.ReLU(inplace=True) | |
self.maxpool = nn.MaxPool2d(kernel_size=3, stride=2, padding=1) | |
self.layer1 = self._make_layer(BasicBlock, 64, 2, stride=1) | |
self.layer2 = self._make_layer(BasicBlock, 128, 2, stride=2) | |
self.layer3 = self._make_layer(BasicBlock, 256, 2, stride=2) | |
self.layer4 = self._make_layer(BasicBlock, 512, 2, stride=2) | |
self.avgpool = nn.AdaptiveAvgPool2d((1, 1)) | |
self.fc = nn.Linear(512 * BasicBlock.expansion, num_classes) | |
def _make_layer(self, block, planes, num_blocks, stride): | |
strides = [stride] + [1] * (num_blocks - 1) | |
layers = [] | |
for stride in strides: | |
layers.append(block(self.in_planes, planes, stride)) | |
self.in_planes = planes * block.expansion | |
return nn.Sequential(*layers) | |
def forward(self, x): | |
out = self.relu(self.bn |