概述
最近想学习下Mark R-CNN模型做实例分割,想快速体验下,所以暂时不想花时间去看源码的具体实现细节.RCNN系列模型介绍推荐看这篇博客目标检测模型概述.里面主要介绍了目标检测模型的发展.后来看到一篇博客链接看到较新版本的torchvision里面有现成的Mask R-CNN模型可以直接使用,且程序里面注释有demo教你快速使用.因此自己就马上体验下啦(文尾有全部代码,utils.py,transforms.py,coco_eval.py,engine.py,coco_utils.py从这里下载链接).
本机环境:系统ubuntu16.4,GTX 960,cuda10.0,cudnn7.6,pytorch1.2.0,torchvision0.4.0
测试预训练模型
导入相应包
import os
import torch
import numpy as np
import torch.utils.data
from PIL import Image
import torchvision
from torchvision.models.detection.faster_rcnn import FastRCNNPredictor
from torchvision.models.detection.mask_rcnn import MaskRCNNPredictor
import utils
import transforms as T
from engine import train_one_epoch, evaluate
import cv2
from random import shuffle
为了后面的检测结果显示,定义coco数据集的80类对象对应的显示颜色
class_names = ['BG', 'person', 'bicycle', 'car', 'motorcycle', 'airplane',
'bus', 'train', 'truck', 'boat', 'traffic light',
'fire hydrant', 'stop sign', 'parking meter', 'bench', 'bird',
'cat', 'dog', 'horse', 'sheep', 'cow', 'elephant', 'bear',
'zebra', 'giraffe', 'backpack', 'umbrella', 'handbag', 'tie',
'suitcase', 'frisbee', 'skis', 'snowboard', 'sports ball',
'kite', 'baseball bat', 'baseball glove', 'skateboard',
'surfboard', 'tennis racket', 'bottle', 'wine glass', 'cup',
'fork', 'knife', 'spoon', 'bowl', 'banana', 'apple',
'sandwich', 'orange', 'broccoli', 'carrot', 'hot dog', 'pizza',
'donut', 'cake', 'chair', 'couch', 'potted plant', 'bed',
'dining table', 'toilet', 'tv', 'laptop', 'mouse', 'remote',
'keyboard', 'cell phone', 'microwave', 'oven', 'toaster',
'sink', 'refrigerator', 'book', 'clock', 'vase', 'scissors',
'teddy bear', 'hair drier', 'toothbrush']
colors=[[np.random.randint(0,255),np.random.randint(0,255),np.random.randint(0,255)]for i in range(100)]
#为了最终实例分割显示明显,定义常见类别为深色
colors[1]=[255,0,0] #person
colors[2]=[0,255,0] #bicycle
colors[3]=[0,0,255] #car
colors[4]=[255,255,0]#motorcycle
直接用model=torchvision.models.detection.maskrcnn_resnet50_fpn(pretrained=True)载入预训练模型,这个函数会下载模型文件的,网速嘛…自己FQ吧.模型输出为一个列表,每个元素是每张图的检测结果–dict,由masks,labels,scores三个key得到最终结果.demo程序如下(分数阈值取的0.5).
def demo():
img_dir='/home/zyk/dataset/PennFudanPed/PNGImages'
#img_dir='/home/zyk/datasets/PennFudanPed/PNGImages'
device = torch.device('cuda') if torch.cuda.is_available() else torch.device('cpu')
# load an instance segmentation model pre-trained on COCO
model = torchvision.models.detection.maskrcnn_resnet50_fpn(pretrained=True)
model.to(device)
imgs=os.listdir(img_dir)
shuffle(imgs)
for i in range(50):
imgsrc=cv2.imread(os.path.join(img_dir,imgs[i]))
all_cls_mask_color = np.zeros_like(imgsrc)
all_cls_mask_index=np.zeros_like(imgsrc)
img = imgsrc / 255.
img=np.transpose(img, (2, 0, 1))
img=torch.tensor(img,dtype=torch.float)
# put the model in evaluation mode
model.eval()
with torch.no_grad():
prediction = model([img.to(device)])
#print(prediction)#可以简单传入一个tensor,看看输出返回结果
scores =prediction[0]['scores']
for idx,score in enumerate(scores):
if score>0.5:
mask=prediction[0]['masks'][idx][0].cpu().numpy()
mask=mask>0.5
cls_id=prediction[0]['labels'][idx].item()
all_cls_mask_color[mask]=colors[cls_id]
all_cls_mask_index[mask]=1
img_weight=cv2.addWeighted(imgsrc,0.4,all_cls_mask_color,0.6,0)#线性混合
all_mask=all_cls_mask_index==1
result=np.copy(imgsrc)
result[all_mask]=img_weight[all_mask] #只取mask的混合部分
union = np.concatenate((imgsrc,result),axis=1)
cv2.imshow('',union)
cv2.waitKey(0)
#cv2.imwrite(os.path.join('./results',imgs[i]),union)
运行一下,效果还是很理想的(注意下图背包,交通灯都是分割出来了,只是颜色比较浅)
训练自己的数据
制作数据集(数据集格式暂时没去看),本文上方链接的博文里面有一个小的数据集链接.暂时用这个数据集测试模型先.
不改变backbone的情况下,修改模型输出为自己的输出,代码如下
def get_instance_segmentation_model(num_classes):
# load an instance segmentation model pre-trained on COCO
model = torchvision.models.detection.maskrcnn_resnet50_fpn(pretrained=True)
# 把模型打印出来,按照名字,输入输出更换backbone,或者改变输出
# get the number of input features for the classifier
in_features = model.roi_heads.box_predictor.cls_score.in_features
# replace the pre-trained head with a new one
model.roi_heads.box_predictor = FastRCNNPredictor(in_features, num_classes)
# now get the number of input features for the mask classifier
in_features_mask = model.roi_heads.mask_predictor.conv5_mask.in_channels
hidden_layer = 256
# and replace the mask predictor with a new one
model.roi_heads.mask_predictor = MaskRCNNPredictor(in_features_mask,
hidden_layer,
num_classes)
如果需要更换backbone,关键在于构建FPN层,pytorch里面也有快速构建FPN层的demo.我自己把backbone修改为resnet18,代码如下
class backbone_body(torch.nn.ModuleDict):
def __init__(self,layers,return_layers):
super().__init__(layers)
self.return_layers = return_layers
def forward(self, x):
out = OrderedDict()
for name, module in self.named_children():
x = module(x)
if name in self.return_layers:
out_name = self.return_layers[name]
out[out_name] = x
return out
class BackboneFPN(torch.nn.Sequential):
def __init__(self, body, fpn, out_channels):
super(BackboneFPN, self).__init__(OrderedDict(
[("body", body), ("fpn", fpn)]))
self.out_channels = out_channels
#代码都是从pytorch源码里面整理出来的
def maskrcnn_resnet18_fpn(num_classes):
src_backbone=torchvision.models.resnet18(pretrained=True)
#去掉后面的全连接层
return_layers = {'layer1': 0, 'layer2': 1, 'layer3': 2, 'layer4': 3}
if not set(return_layers).issubset([name for name, _ in src_backbone.named_children()]):
raise ValueError("return_layers are not present in model")
orig_return_layers = return_layers
return_layers = {k: v for k, v in return_layers.items()} #复制一份
layers = OrderedDict()
for name, module in src_backbone.named_children():
layers[name] = module
if name in return_layers:
del return_layers[name]
if not return_layers:
break
backbone_module=backbone_body(layers,orig_return_layers) #得到去掉池化、全连接层的模型
#FPN层,resnet18 layer4 chanels为 512,fpn顶层512/8
in_channels_stage2 = 64
in_channels_list = [
in_channels_stage2,
in_channels_stage2 * 2,
in_channels_stage2 * 4,
in_channels_stage2 * 8,
]
out_channels = 64
fpn = FeaturePyramidNetwork(
in_channels_list=in_channels_list,
out_channels=out_channels,
extra_blocks=LastLevelMaxPool(),
)
backbone_fpn=BackboneFPN(backbone_module,fpn,out_channels)
model = MaskRCNN(backbone_fpn, num_classes)
return model
其余部分就是制作数据集了和训练了.若只是简单训练自己的数据,直接看最后的全部代码就可以了.我后面还会详细看pytorch里面的源码配合相关论文详细了解Mask R-CNN的网络结构和实现细节的(强迫症),后续再详细记录下.
demo和训练自己的数据全部代码如下
# -*- coding:utf-8 -*-
# @Author : ZYK
import os
import torch
import numpy as np
import torch.utils.data
import utils
import transforms as T
import cv2
from engine import train_one_epoch, evaluate
from PIL import Image
from random import shuffle
from collections import OrderedDict
import torchvision
from torchvision.models.detection.faster_rcnn import FastRCNNPredictor
from torchvision.models.detection.mask_rcnn import MaskRCNNPredictor,MaskRCNN
from torchvision.ops.feature_pyramid_network import FeaturePyramidNetwork, LastLevelMaxPool
class_names = ['BG', 'person', 'bicycle', 'car', 'motorcycle', 'airplane',
'bus', 'train', 'truck', 'boat', 'traffic light',
'fire hydrant', 'stop sign', 'parking meter', 'bench', 'bird',
'cat', 'dog', 'horse', 'sheep', 'cow', 'elephant', 'bear',
'zebra', 'giraffe', 'backpack', 'umbrella', 'handbag', 'tie',
'suitcase', 'frisbee', 'skis', 'snowboard', 'sports ball',
'kite', 'baseball bat', 'baseball glove', 'skateboard',
'surfboard', 'tennis racket', 'bottle', 'wine glass', 'cup',
'fork', 'knife', 'spoon', 'bowl', 'banana', 'apple',
'sandwich', 'orange', 'broccoli', 'carrot', 'hot dog', 'pizza',
'donut', 'cake', 'chair', 'couch', 'potted plant', 'bed',
'dining table', 'toilet', 'tv', 'laptop', 'mouse', 'remote',
'keyboard', 'cell phone', 'microwave', 'oven', 'toaster',
'sink', 'refrigerator', 'book', 'clock', 'vase', 'scissors',
'teddy bear', 'hair drier', 'toothbrush']
colors=[[np.random.randint(0,255),np.random.randint(0,255),np.random.randint(0,255)]for i in range(100)]
#为了最终实例分割显示明显,定义常见类别为深色
colors[1]=[255,0,0] #person
colors[2]=[0,255,0] #bicycle
colors[3]=[0,0,255] #car
colors[4]=[255,255,0]#motorcycle
def demo():
img_dir='/home/zyk/datasets/PennFudanPed/PNGImages'
#img_dir='/home/zyk/datasets/PennFudanPed/PNGImages'
device = torch.device('cuda') if torch.cuda.is_available() else torch.device('cpu')
# load an instance segmentation model pre-trained on COCO
model = torchvision.models.detection.maskrcnn_resnet50_fpn(pretrained=True)
model.to(device)
imgs=os.listdir(img_dir)
shuffle(imgs)
for i in range(50):
imgsrc=cv2.imread(os.path.join(img_dir,imgs[i]))
all_cls_mask_color = np.zeros_like(imgsrc)
all_cls_mask_index=np.zeros_like(imgsrc)
img = imgsrc / 255.
img=np.transpose(img, (2, 0, 1))
img=torch.tensor(img,dtype=torch.float)
# put the model in evaluation mode
model.eval()
with torch.no_grad():
prediction = model([img.to(device)])
#print(prediction)#可以简单传入一个tensor,看看输出返回结果
scores =prediction[0]['scores']
for idx,score in enumerate(scores):
if score>0.5:
mask=prediction[0]['masks'][idx][0].cpu().numpy()
mask=mask>0.5
cls_id=prediction[0]['labels'][idx].item()
all_cls_mask_color[mask]=colors[cls_id]
all_cls_mask_index[mask]=1
img_weight=cv2.addWeighted(imgsrc,0.4,all_cls_mask_color,0.6,0)#线性混合
all_mask=all_cls_mask_index==1
result=np.copy(imgsrc)
result[all_mask]=img_weight[all_mask] #只取mask的混合部分
union = np.concatenate((imgsrc,result),axis=1)
cv2.imshow('',union)
cv2.waitKey(0)
#cv2.imwrite(os.path.join('./results',imgs[i]),union)
dataset_dir='/home/zyk/datasets/PennFudanPed'
class PennFudanDataset(torch.utils.data.Dataset):
def __init__(self, root, transforms=None):
self.root = root
self.transforms = transforms
# load all image files, sorting them to ensure that they are aligned
self.imgs = list(sorted(os.listdir(os.path.join(root, 'PNGImages'))))
self.masks = list(sorted(os.listdir(os.path.join(root, 'PedMasks'))))
def __getitem__(self, idx):
# load images ad masks
img_path = os.path.join(self.root, "PNGImages", self.imgs[idx])
mask_path = os.path.join(self.root, "PedMasks", self.masks[idx])
img = Image.open(img_path).convert("RGB")
# note that we haven't converted the mask to RGB,
# because each color corresponds to a different instance with 0 being background
mask = Image.open(mask_path)
mask = np.array(mask)
# instances are encoded as different colors
obj_ids = np.unique(mask)
# first id is the background, so remove it
obj_ids = obj_ids[1:]
# split the color-encoded mask into a set of binary masks
masks = mask == obj_ids[:, None, None]
# get bounding box coordinates for each mask
num_objs = len(obj_ids)
boxes = []
for i in range(num_objs):
pos = np.where(masks[i])
xmin = np.min(pos[1])
xmax = np.max(pos[1])
ymin = np.min(pos[0])
ymax = np.max(pos[0])
boxes.append([xmin, ymin, xmax, ymax])
boxes = torch.as_tensor(boxes, dtype=torch.float32)
# there is only one class
labels = torch.ones((num_objs,), dtype=torch.int64)
masks = torch.as_tensor(masks, dtype=torch.uint8)
image_id = torch.tensor([idx])
area = (boxes[:, 3] - boxes[:, 1]) * (boxes[:, 2] - boxes[:, 0])
# suppose all instances are not crowd
iscrowd = torch.zeros((num_objs,), dtype=torch.int64)
target = {}
target["boxes"] = boxes
target["labels"] = labels
target["masks"] = masks
target["image_id"] = image_id
target["area"] = area
target["iscrowd"] = iscrowd
if self.transforms is not None:
img, target = self.transforms(img, target)
return img, target
def __len__(self):
return len(self.imgs)
def get_transform(train):
transforms = []
# converts the image, a PIL image, into a PyTorch Tensor
transforms.append(T.ToTensor())
if train:
# during training, randomly flip the training images
# and ground-truth for data augmentation
transforms.append(T.RandomHorizontalFlip(0.5)) #注意没有resize
return T.Compose(transforms)
class backbone_body(torch.nn.ModuleDict):
def __init__(self,layers,return_layers):
super().__init__(layers)
self.return_layers = return_layers
def forward(self, x):
out = OrderedDict()
for name, module in self.named_children():
x = module(x)
if name in self.return_layers:
out_name = self.return_layers[name]
out[out_name] = x
return out
class BackboneFPN(torch.nn.Sequential):
def __init__(self, body, fpn, out_channels):
super(BackboneFPN, self).__init__(OrderedDict(
[("body", body), ("fpn", fpn)]))
self.out_channels = out_channels
#代码都是从pytorch源码里面整理出来的
def maskrcnn_resnet18_fpn(num_classes):
src_backbone=torchvision.models.resnet18(pretrained=True)
#去掉后面的全连接层
return_layers = {'layer1': 0, 'layer2': 1, 'layer3': 2, 'layer4': 3}
if not set(return_layers).issubset([name for name, _ in src_backbone.named_children()]):
raise ValueError("return_layers are not present in model")
orig_return_layers = return_layers
return_layers = {k: v for k, v in return_layers.items()} #复制一份
layers = OrderedDict()
for name, module in src_backbone.named_children():
layers[name] = module
if name in return_layers:
del return_layers[name]
if not return_layers:
break
backbone_module=backbone_body(layers,orig_return_layers) #得到去掉池化、全连接层的模型
#FPN层,resnet18 layer4 chanels为 512,fpn顶层512/8
in_channels_stage2 = 64
in_channels_list = [
in_channels_stage2,
in_channels_stage2 * 2,
in_channels_stage2 * 4,
in_channels_stage2 * 8,
]
out_channels = 64
fpn = FeaturePyramidNetwork(
in_channels_list=in_channels_list,
out_channels=out_channels,
extra_blocks=LastLevelMaxPool(),
)
backbone_fpn=BackboneFPN(backbone_module,fpn,out_channels)
model = MaskRCNN(backbone_fpn, num_classes)
return model
def get_instance_segmentation_model(num_classes):
# load an instance segmentation model pre-trained on COCO
model = torchvision.models.detection.maskrcnn_resnet50_fpn(pretrained=True)
# 把模型打印出来,按照名字,输入输出更换backbone,或者改变输出
# get the number of input features for the classifier
in_features = model.roi_heads.box_predictor.cls_score.in_features
# replace the pre-trained head with a new one
model.roi_heads.box_predictor = FastRCNNPredictor(in_features, num_classes)
# now get the number of input features for the mask classifier
in_features_mask = model.roi_heads.mask_predictor.conv5_mask.in_channels
hidden_layer = 256
# and replace the mask predictor with a new one
model.roi_heads.mask_predictor = MaskRCNNPredictor(in_features_mask,
hidden_layer,
num_classes)
return model
def train_mydata():
# use the PennFudan dataset and defined transformations
dataset = PennFudanDataset(dataset_dir, get_transform(train=True))
dataset_test = PennFudanDataset(dataset_dir, get_transform(train=False))
# split the dataset in train and test set
torch.manual_seed(1)
indices = torch.randperm(len(dataset)).tolist()
dataset = torch.utils.data.Subset(dataset, indices[:-50])
dataset_test = torch.utils.data.Subset(dataset_test, indices[-50:])
# define training and validation data loaders
data_loader = torch.utils.data.DataLoader(
dataset, batch_size=2, shuffle=True, num_workers=4,
collate_fn=utils.collate_fn)
data_loader_test = torch.utils.data.DataLoader(
dataset_test, batch_size=1, shuffle=False, num_workers=4,
collate_fn=utils.collate_fn)
device = torch.device('cuda') if torch.cuda.is_available() else torch.device('cpu')
# the dataset has two classes only - background and person
num_classes = 2
#model = get_instance_segmentation_model(num_classes)#只改输出类别数
model = maskrcnn_resnet18_fpn(num_classes) #更换backbone
# move model to the right device
model.to(device)
# construct an optimizer
params = [p for p in model.parameters() if p.requires_grad]
optimizer = torch.optim.SGD(params, lr=0.005,
momentum=0.9, weight_decay=0.0005)
# the learning rate scheduler decreases the learning rate by 10x every 3 epochs
lr_scheduler = torch.optim.lr_scheduler.StepLR(optimizer,
step_size=3,
gamma=0.1)
# training
print('---------------------------start train--------------------------')
num_epochs = 10
for epoch in range(num_epochs):
# train for one epoch, printing every 10 iterations
train_one_epoch(model, optimizer, data_loader, device, epoch, print_freq=10)
# update the learning rate
lr_scheduler.step()
# evaluate on the test dataset
evaluate(model, data_loader_test, device=device)
print('---------------------------finish train------------------------')
# 测试
for i in range(10):
img, _ = dataset_test[i]
# put the model in evaluation mode
model.eval()
with torch.no_grad():
prediction = model([img.to(device)])
src = img.mul(255).permute(1, 2, 0).byte().numpy()
result = prediction[0]['masks'][0, 0].mul(255).byte().cpu().numpy()
result = np.expand_dims(result, -1).repeat(3, axis=-1)
result = cv2.addWeighted(src, 0.5, result, 0.5, 0)
cv2.imshow("result", result)
cv2.waitKey(0)
if __name__=='__main__':
#demo()
train_mydata()