为了方便实验对比结果,先在将visdrone的验证集可视化出来,脚本是git别人的,实测可行,分享给需要的人吧。
代码是根据xml标签可视化的,需要将visdrone的txt标签转成xml,可参考我另一篇文章:
import os
import os.path
import numpy as np
import xml.etree.ElementTree as xmlET
from PIL import Image, ImageDraw
#'1': 'people', '2': 'people','3': 'bicycle', '4': 'car', '5': 'car',
# 6':'others','7':'others','8':'others','9':'others','10': 'motor','11':'others'
classes = ('__background__', # always index 0
'ignored regions','pedestrian', 'people','bicycle','car','van','truck','tricycle','awning-tricycle',
'bus','motor','others')
# 把下面的路径改为自己的路径即可
file_path_img = r'D:\object_detection_data\datacovert\VisDrone2019-DET-val\images'
file_path_xml = r'D:\object_detection_data\datacovert\VisDrone2019-DET-val\Annotations_XML'
save_file_path = r'D:\object_detection_data\datacovert\VisDrone2019-DET-val\image_ann_show'
pathDir = os.listdir(file_path_xml)
for idx in range(len(pathDir)):
filename = pathDir[idx]
tree = xmlET.parse(os.path.join(file_path_xml, filename))
objs = tree.findall('object')
num_objs = len(objs)
boxes = np.zeros((num_objs, 5), dtype=np.uint16)
for ix, obj in enumerate(objs):
bbox = obj.find('bndbox')
# Make pixel indexes 0-based
x1 = float(bbox.find('xmin').text)
y1 = float(bbox.find('ymin').text)
x2 = float(bbox.find('xmax').text)
y2 = float(bbox.find('ymax').text)
cla = obj.find('name').text
label = classes.index(cla)
boxes[ix, 0:4] = [x1, y1, x2, y2]
boxes[ix, 4] = label
image_name = os.path.splitext(filename)[0]
img = Image.open(os.path.join(file_path_img, image_name + '.jpg'))
draw = ImageDraw.Draw(img)
for ix in range(len(boxes)):
xmin = int(boxes[ix, 0])
ymin = int(boxes[ix, 1])
xmax = int(boxes[ix, 2])
ymax = int(boxes[ix, 3])
draw.rectangle([xmin, ymin, xmax, ymax], outline=(255, 0, 0))
draw.text([xmin, ymin], classes[boxes[ix, 4]], (255, 0, 0))
img.save(os.path.join(save_file_path, image_name + '.png'))
如代码所示只需要修改三个路径即可
可视化结果:
网上这方面的代码较少,找到了两三个不同版本,这个代码可视化出来的效果比较好。