[bevformer渲染可视化] GT与PRED进行iou匹配,输出漏检误检

可视化代码:


代码使用方法:
1.复制代码全文到BEVFormer/tools/analysis_tools/visual.py文件中

2.修改bevformer_results = mmcv.load('test/bevformer_tiny/Sat_Jul__6_12_03_22_2024/pts_bbox/results_nusc.json')路径

3.可修改iou值,进行不同iou匹配输出不同程度的匹配结果

 # 仅处理相机数据
          if sensor_modality == 'camera':
              # 创建一个预测框列表,只有预测分数大于 0.5 的记录才会被包含在内
              boxes = [Box(record['translation'], record['size'], Quaternion(record['rotation']),
                           name=record['detection_name'], token='predicted') for record in
                       pred_data['results'][sample_token] if record['detection_score'] > 0.5]

4.运行命令

python tools/analysis_tools/visual.py
# Based on https://github.com/nutonomy/nuscenes-devkit
# ---------------------------------------------
#  Modified by Zhiqi Li
# ---------------------------------------------

import mmcv
from nuscenes.nuscenes import NuScenes
from PIL import Image
from nuscenes.utils.geometry_utils import view_points, box_in_image, BoxVisibility, transform_matrix
from typing import Tuple, List, Iterable
import matplotlib.pyplot as plt
import numpy as np
from PIL import Image
from matplotlib import rcParams
from matplotlib.axes import Axes
from pyquaternion import Quaternion
from PIL import Image
from matplotlib import rcParams
from matplotlib.axes import Axes
from pyquaternion import Quaternion
from tqdm import tqdm
from nuscenes.utils.data_classes import LidarPointCloud, RadarPointCloud, Box
from nuscenes.utils.geometry_utils import view_points, box_in_image, BoxVisibility, transform_matrix
from nuscenes.eval.common.data_classes import EvalBoxes, EvalBox
from nuscenes.eval.detection.data_classes import DetectionBox
from nuscenes.eval.detection.utils import category_to_detection_name
from nuscenes.eval.detection.render import visualize_sample
import pandas as pd
from openpyxl import Workbook

import os
import cv2


cams = ['CAM_FRONT',
 'CAM_FRONT_RIGHT',
 'CAM_BACK_RIGHT',
 'CAM_BACK',
 'CAM_BACK_LEFT',
 'CAM_FRONT_LEFT']

import numpy as np
import matplotlib.pyplot as plt
from nuscenes.utils.data_classes import LidarPointCloud, RadarPointCloud, Box
from PIL import Image
from matplotlib import rcParams


def render_annotation(
        anntoken: str,
        margin: float = 10,
        view: np.ndarray = np.eye(4),
        box_vis_level: BoxVisibility = BoxVisibility.ANY,
        out_path: str = 'render.png',
        extra_info: bool = False) -> None:
    """
    Render selected annotation.
    :param anntoken: Sample_annotation token.
    :param margin: How many meters in each direction to include in LIDAR view.
    :param view: LIDAR view point.
    :param box_vis_level: If sample_data is an image, this sets required visibility for boxes.
    :param out_path: Optional path to save the rendered figure to disk.
    :param extra_info: Whether to render extra information below camera view.
    """
    ann_record = nusc.get('sample_annotation', anntoken)
    sample_record = nusc.get('sample', ann_record['sample_token'])
    assert 'LIDAR_TOP' in sample_record['data'].keys(), 'Error: No LIDAR_TOP in data, unable to render.'

    # Figure out which camera the object is fully visible in (this may return nothing).
    boxes, cam = [], []
    cams = [key for key in sample_record['data'].keys() if 'CAM' in key]
    all_bboxes = []
    select_cams = []
    for cam in cams:
        _, boxes, _ = nusc.get_sample_data(sample_record['data'][cam], box_vis_level=box_vis_level,
                                           selected_anntokens=[anntoken])
        if len(boxes) > 0:
            all_bboxes.append(boxes)
            select_cams.append(cam)
            # We found an image that matches. Let's abort.
    # assert len(boxes) > 0, 'Error: Could not find image where annotation is visible. ' \
    #                      'Try using e.g. BoxVisibility.ANY.'
    # assert len(boxes) < 2, 'Error: Found multiple annotations. Something is wrong!'

    num_cam = len(all_bboxes)

    fig, axes = plt.subplots(1, num_cam + 1, figsize=(18, 9))
    select_cams = [sample_record['data'][cam] for cam in select_cams]
    print('bbox in cams:', select_cams)
    # Plot LIDAR view.
    lidar = sample_record['data']['LIDAR_TOP']
    data_path, boxes, camera_intrinsic = nusc.get_sample_data(lidar, selected_anntokens=[anntoken])
    LidarPointCloud.from_file(data_path).render_height(axes[0], view=view)
    for box in boxes:
        c = np.array(get_color(box.name)) / 255.0
        box.render(axes[0], view=view, colors=(c, c, c))
        corners = view_points(boxes[0].corners(), view, False)[:2, :]
        axes[0].set_xlim([np.min(corners[0, :]) - margin, np.max(corners[0, :]) + margin])
        axes[0].set_ylim([np.min(corners[1, :]) - margin, np.max(corners[1, :]) + margin])
        axes[0].axis('off')
        axes[0].set_aspect('equal')

    # Plot CAMERA view.
    for i in range(1, num_cam + 1):
        cam = select_cams[i - 1]
        data_path, boxes, camera_intrinsic = nusc.get_sample_data(cam, selected_anntokens=[anntoken])
        im = Image.open(data_path)
        axes[i].imshow(im)
        axes[i].set_title(nusc.get('sample_data', cam)['channel'])
        axes[i].axis('off')
        axes[i].set_aspect('equal')
        for box in boxes:
            c = np.array(get_color(box.name)) / 255.0
            box.render(axes[i], view=camera_intrinsic, normalize=True, colors=(c, c, c))

        # Print extra information about the annotation below the camera view.
        axes[i].set_xlim(0, im.size[0])
        axes[i].set_ylim(im.size[1], 0)

    if extra_info:
        rcParams['font.family'] = 'monospace'

        w, l, h = ann_record['size']
        category = ann_record['category_name']
        lidar_points = ann_record['num_lidar_pts']
        radar_points = ann_record['num_radar_pts']

        sample_data_record = nusc.get('sample_data', sample_record['data']['LIDAR_TOP'])
        pose_record = nusc.get('ego_pose', sample_data_record['ego_pose_token'])
        dist = np.linalg.norm(np.array(pose_record['translation']) - np.array(ann_record['translation']))

        information = ' \n'.join(['category: {}'.format(category),
                                  '',
                                  '# lidar points: {0:>4}'.format(lidar_points),
                                  '# radar points: {0:>4}'.format(radar_points),
                                  '',
                                  'distance: {:>7.3f}m'.format(dist),
                                  '',
                                  'width:  {:>7.3f}m'.format(w),
                                  'length: {:>7.3f}m'.format(l),
                                  'height: {:>7.3f}m'.format(h)])

        plt.annotate(information, (0, 0), (0, -20), xycoords='axes fraction', textcoords='offset points', va='top')

    if out_path is not None:
        plt.savefig(out_path)



def get_sample_data(sample_data_token: str,
                    box_vis_level: BoxVisibility = BoxVisibility.ANY,
                    selected_anntokens=None,
                    use_flat_vehicle_coordinates: bool = False):
    """
    Returns the data path as well as all annotations related to that sample_data.
    Note that the boxes are transformed into the current sensor's coordinate frame.
    :param sample_data_token: Sample_data token.
    :param box_vis_level: If sample_data is an image, this sets required visibility for boxes.
    :param selected_anntokens: If provided only return the selected annotation.
    :param use_flat_vehicle_coordinates: Instead of the current sensor's coordinate frame, use ego frame which is
                                         aligned to z-plane in the world.
    :return: (data_path, boxes, camera_intrinsic <np.array: 3, 3>)
    """

    # Retrieve sensor & pose records
    sd_record = nusc.get('sample_data', sample_data_token)
    cs_record = nusc.get('calibrated_sensor', sd_record['calibrated_sensor_token'])
    sensor_record = nusc.get('sensor', cs_record['sensor_token'])
    pose_record = nusc.get('ego_pose', sd_record['ego_pose_token'])

    data_path = nusc.get_sample_data_path(sample_data_token)

    if sensor_record['modality'] == 'camera':
        cam_intrinsic = np.array(cs_record['camera_intrinsic'])
        imsize = (sd_record['width'], sd_record['height'])
    else:
        cam_intrinsic = None
        imsize = None

    # Retrieve all sample annotations and map to sensor coordinate system.
    if selected_anntokens is not None:
        boxes = list(map(nusc.get_box, selected_anntokens))
    else:
        boxes = nusc.get_boxes(sample_data_token)

    # Make list of Box objects including coord system transforms.
    box_list = []
    for box in boxes:
        if use_flat_vehicle_coordinates:
            # Move box to ego vehicle coord system parallel to world z plane.
            yaw = Quaternion(pose_record['rotation']).yaw_pitch_roll[0]
            box.translate(-np.array(pose_record['translation']))
            box.rotate(Quaternion(scalar=np.cos(yaw / 2), vector=[0, 0, np.sin(yaw / 2)]).inverse)
        else:
            # Move box to ego vehicle coord system.
            box.translate(-np.array(pose_record['translation']))
            box.rotate(Quaternion(pose_record['rotation']).inverse)

            #  Move box to sensor coord system.
            box.translate(-np.array(cs_record['translation']))
            box.rotate(Quaternion(cs_record['rotation']).inverse)

        if sensor_record['modality'] == 'camera' and not \
                box_in_image(box, cam_intrinsic, imsize, vis_level=box_vis_level):
            continue

        box_list.append(box)

    return data_path, box_list, cam_intrinsic



def get_predicted_data(sample_data_token: str,
                       box_vis_level: BoxVisibility = BoxVisibility.ANY,
                       selected_anntokens=None,
                       use_flat_vehicle_coordinates: bool = False,
                       pred_anns=None
                       ):
    """
    Returns the data path as well as all annotations related to that sample_data.
    Note that the boxes are transformed into the current sensor's coordinate frame.
    :param sample_data_token: Sample_data token.
    :param box_vis_level: If sample_data is an image, this sets required visibility for boxes.
    :param selected_anntokens: If provided only return the selected annotation.
    :param use_flat_vehicle_coordinates: Instead of the current sensor's coordinate frame, use ego frame which is
                                         aligned to z-plane in the world.
    :return: (data_path, boxes, camera_intrinsic <np.array: 3, 3>)
    """

    # Retrieve sensor & pose records
    sd_record = nusc.get('sample_data', sample_data_token)
    cs_record = nusc.get('calibrated_sensor', sd_record['calibrated_sensor_token'])
    sensor_record = nusc.get('sensor', cs_record['sensor_token'])
    pose_record = nusc.get('ego_pose', sd_record['ego_pose_token'])

    data_path = nusc.get_sample_data_path(sample_data_token)

    if sensor_record['modality'] == 'camera':
        cam_intrinsic = np.array(cs_record['camera_intrinsic'])
        imsize = (sd_record['width'], sd_record['height'])
    else:
        cam_intrinsic = None
        imsize = None

    # Retrieve all sample annotations and map to sensor coordinate system.
    # if selected_anntokens is not None:
    #    boxes = list(map(nusc.get_box, selected_anntokens))
    # else:
    #    boxes = nusc.get_boxes(sample_data_token)
    boxes = pred_anns
    # Make list of Box objects including coord system transforms.
    box_list = []
    for box in boxes:
        if use_flat_vehicle_coordinates:
            # Move box to ego vehicle coord system parallel to world z plane.
            yaw = Quaternion(pose_record['rotation']).yaw_pitch_roll[0]
            box.translate(-np.array(pose_record['translation']))
            box.rotate(Quaternion(scalar=np.cos(yaw / 2), vector=[0, 0, np.sin(yaw / 2)]).inverse)
        else:
            # Move box to ego vehicle coord system.
            box.translate(-np.array(pose_record['translation']))
            box.rotate(Quaternion(pose_record['rotation']).inverse)

            #  Move box to sensor coord system.
            box.translate(-np.array(cs_record['translation']))
            box.rotate(Quaternion(cs_record['rotation']).inverse)

        if sensor_record['modality'] == 'camera' and not \
                box_in_image(box, cam_intrinsic, imsize, vis_level=box_vis_level):
            continue
        box_list.append(box)

    return data_path, box_list, cam_intrinsic




def lidiar_render(sample_token, data,out_path=None):
    bbox_gt_list = []
    bbox_pred_list = []
    anns = nusc.get('sample', sample_token)['anns']
    for ann in anns:
        content = nusc.get('sample_annotation', ann)
        try:
            bbox_gt_list.append(DetectionBox(
                sample_token=content['sample_token'],
                translation=tuple(content['translation']),
                size=tuple(content['size']),
                rotation=tuple(content['rotation']),
                velocity=nusc.box_velocity(content['token'])[:2],
                ego_translation=(0.0, 0.0, 0.0) if 'ego_translation' not in content
                else tuple(content['ego_translation']),
                num_pts=-1 if 'num_pts' not in content else int(content['num_pts']),
                detection_name=category_to_detection_name(content['category_name']),
                detection_score=-1.0 if 'detection_score' not in content else float(content['detection_score']),
                attribute_name=''))
        except:
            pass

    bbox_anns = data['results'][sample_token]
    for content in bbox_anns:
        bbox_pred_list.append(DetectionBox(
            sample_token=content['sample_token'],
            translation=tuple(content['translation']),
            size=tuple(content['size']),
            rotation=tuple(content['rotation']),
            velocity=tuple(content['velocity']),
            ego_translation=(0.0, 0.0, 0.0) if 'ego_translation' not in content
            else tuple(content['ego_translation']),
            num_pts=-1 if 'num_pts' not in content else int(content['num_pts']),
            detection_name=content['detection_name'],
            detection_score=-1.0 if 'detection_score' not in content else float(content['detection_score']),
            attribute_name=content['attribute_name']))
    gt_annotations = EvalBoxes()
    pred_annotations = EvalBoxes()
    gt_annotations.add_boxes(sample_token, bbox_gt_list)
    pred_annotations.add_boxes(sample_token, bbox_pred_list)
    print('green is ground truth')
    print('blue is the predited result')
    visualize_sample(nusc, sample_token, gt_annotations, pred_annotations, savepath=out_path+'_bev')


def get_color(category_name: str):
    """
    Provides the default colors based on the category names.
    This method works for the general nuScenes categories, as well as the nuScenes detection categories.
    """
    a = ['noise', 'animal', 'human.pedestrian.adult', 'human.pedestrian.child', 'human.pedestrian.construction_worker',
     'human.pedestrian.personal_mobility', 'human.pedestrian.police_officer', 'human.pedestrian.stroller',
     'human.pedestrian.wheelchair', 'movable_object.barrier', 'movable_object.debris',
     'movable_object.pushable_pullable', 'movable_object.trafficcone', 'static_object.bicycle_rack', 'vehicle.bicycle',
     'vehicle.bus.bendy', 'vehicle.bus.rigid', 'vehicle.car', 'vehicle.construction', 'vehicle.emergency.ambulance',
     'vehicle.emergency.police', 'vehicle.motorcycle', 'vehicle.trailer', 'vehicle.truck', 'flat.driveable_surface',
     'flat.other', 'flat.sidewalk', 'flat.terrain', 'static.manmade', 'static.other', 'static.vegetation',
     'vehicle.ego']
    class_names = [
        'car', 'truck', 'construction_vehicle', 'bus', 'trailer', 'barrier',
        'motorcycle', 'bicycle', 'pedestrian', 'traffic_cone'
    ]
    #print(category_name)
    if category_name == 'bicycle':
        return nusc.colormap['vehicle.bicycle']
    elif category_name == 'construction_vehicle':
        return nusc.colormap['vehicle.construction']
    elif category_name == 'traffic_cone':
        return nusc.colormap['movable_object.trafficcone']

    for key in nusc.colormap.keys():
        if category_name in key:
            return nusc.colormap[key]
    return [0, 0, 0]


def crop_and_save_boxes(image_path, boxes_2d, output_dir):
    """
    根据2D边界框裁剪图像并保存裁剪后的图像
    """
    # 打开图像
    image = Image.open(image_path)
    
    # 创建输出目录
    if not os.path.exists(output_dir):
        os.makedirs(output_dir)
    
    # 遍历每个边界框并裁剪图像
    for i, corners in enumerate(boxes_2d):
        x_min = max(int(min(corners[:, 0])), 0)
        x_max = min(int(max(corners[:, 0])), image.width)
        y_min = max(int(min(corners[:, 1])), 0)
        y_max = min(int(max(corners[:, 1])), image.height)
        
        cropped_image = image.crop((x_min, y_min, x_max, y_max))
        cropped_image.save(os.path.join(output_dir, f"cropped_box_{i}.png"))

def project_boxes_to_2d(boxes, camera_intrinsic, image_size):
    """
    将3D边界框投影到2D图像平面
    """
    projected_boxes = []
    for box in boxes:
        
        if isinstance(box, Box):
            box_obj = box
        else:
            # 假设 box 是 numpy.ndarray,形状为 (8, 3)
            translation = np.mean(box, axis=0)
            size = np.ptp(box, axis=0)
            rotation = np.eye(3)
            box_obj = Box(translation, size, rotation)
       
        corners_3d = box_obj.corners()  # 获取3D框的角点
 
        # 投影到2D图像平面
        is_in_image = box_in_image(box_obj, camera_intrinsic, image_size)
        if is_in_image:
            corners_2d = view_points(corners_3d, camera_intrinsic, normalize=True)

            projected_boxes.append(corners_2d.T)
    return projected_boxes

# 计算两个边界框的重叠度IOU
def compute_iou(box1, box2):   
        x1, y1, x2, y2 = box1
        x1_p, y1_p, x2_p, y2_p = box2

        # 计算交集的坐标
        xi1 = max(x1, x1_p)
        yi1 = max(y1, y1_p)
        xi2 = min(x2, x2_p)
        yi2 = min(y2, y2_p)
        inter_area = max(0, xi2 - xi1) * max(0, yi2 - yi1)

        # 计算各自的面积
        box1_area = (x2 - x1) * (y2 - y1)
        box2_area = (x2_p - x1_p) * (y2_p - y1_p)

        # 计算交并比
        iou = inter_area / float(box1_area + box2_area - inter_area)
        return iou

# 得到2dbox四个点的值
def get_2dcorners(corners):
    
    xs = [point[0] for point in corners]
    ys = [point[1] for point in corners]
    x1 = min(xs)
    y1 = min(ys)
    x2 = max(xs)
    y2 = max(ys)
    return (x1, y1, x2, y2)


# 找出未匹配的 gt 边界框
def find_unmatched_gt(gt_boxes, pred_boxes, iou_threshold=0.5):
    unmatched_gt = []

    for gt in gt_boxes:
        gt_2dcorners = get_2dcorners(gt)
        matched = False
        for pred in pred_boxes:
            pred_2dcorners = get_2dcorners(pred)
            iou = compute_iou( gt_2dcorners, pred_2dcorners)
            if iou >= iou_threshold:
                matched = True
                break
        if not matched:
            unmatched_gt.append(gt)

    return unmatched_gt

# 找出已匹配的 gt 边界框
def find_matched_gt(gt_boxes, pred_boxes, iou_threshold=0.5):
    matched_gt = []

    for gt in gt_boxes:
        gt_2dcorners = get_2dcorners(gt)
        matched = False
        for pred in pred_boxes:
            pred_2dcorners = get_2dcorners(pred)
            iou = compute_iou( gt_2dcorners, pred_2dcorners)
            if iou >= iou_threshold:
                matched = True
                break
        if matched:
            matched_gt.append(gt)

    return matched_gt

# 计算长宽及面积
def calculate_box_data(box):
    box = get_2dcorners(box)
    x1, y1, x2, y2 = box
    width = x2 - x1
    height = y2 - y1
    area = width * height
    return width, height, area

# 计算2d框相对图像的大小
def calculate_area_ratio(box, image_size):

    W, H = image_size
    
    # 计算2D框的面积
    _ , _, area_box= calculate_box_data(box)
    
    # 计算图片的面积
    area_image = W * H
    
    # 计算比例
    ratio = area_box / area_image
    
    # 转换为百分比
    percentage = ratio * 100
    
    return  percentage


def save_to_excel(data,excel_output_path, sheet_name='Sheet1'):
    """
    将数据保存到 Excel 文件中,如果文件已存在则追加数据。
    """
    
    
    # 如果文件存在,则读取现有数据并追加
    if os.path.exists(excel_output_path):
        # 读取现有的 Excel 文件
        existing_df = pd.read_excel(excel_output_path, sheet_name=sheet_name)
        # 将现有数据与新数据合并
        new_df = pd.DataFrame(data)
        combined_df = pd.concat([existing_df, new_df], ignore_index=True)
    else:
        # 文件不存在,则创建新的 DataFrame
        combined_df = pd.DataFrame(data)
    
    # 将 DataFrame 写入 Excel 文件
    with pd.ExcelWriter(excel_output_path, engine='xlsxwriter') as writer:
        combined_df.to_excel(writer, sheet_name=sheet_name, index=False)  

def render_sample_data(
        sample_token: str,
        with_anns: bool = True,
        box_vis_level: BoxVisibility = BoxVisibility.ANY,
        axes_limit: float = 40,
        ax=None,
        nsweeps: int = 1,
        out_path: str = None,
        underlay_map: bool = True,
        use_flat_vehicle_coordinates: bool = True,
        show_lidarseg: bool = False,
        show_lidarseg_legend: bool = False,
        filter_lidarseg_labels=None,
        lidarseg_preds_bin_path: str = None,
        verbose: bool = True,
        show_panoptic: bool = False,
        pred_data=None,
      ) -> None:
    """
    Render sample data onto axis.
    :param sample_data_token: Sample_data token.
    :param with_anns: Whether to draw box annotations.
    :param box_vis_level: If sample_data is an image, this sets required visibility for boxes.
    :param axes_limit: Axes limit for lidar and radar (measured in meters).
    :param ax: Axes onto which to render.
    :param nsweeps: Number of sweeps for lidar and radar.
    :param out_path: Optional path to save the rendered figure to disk.
    :param underlay_map: When set to true, lidar data is plotted onto the map. This can be slow.
    :param use_flat_vehicle_coordinates: Instead of the current sensor's coordinate frame, use ego frame which is
        aligned to z-plane in the world. Note: Previously this method did not use flat vehicle coordinates, which
        can lead to small errors when the vertical axis of the global frame and lidar are not aligned. The new
        setting is more correct and rotates the plot by ~90 degrees.
    :param show_lidarseg: When set to True, the lidar data is colored with the segmentation labels. When set
        to False, the colors of the lidar data represent the distance from the center of the ego vehicle.
    :param show_lidarseg_legend: Whether to display the legend for the lidarseg labels in the frame.
    :param filter_lidarseg_labels: Only show lidar points which belong to the given list of classes. If None
        or the list is empty, all classes will be displayed.
    :param lidarseg_preds_bin_path: A path to the .bin file which contains the user's lidar segmentation
                                    predictions for the sample.
    :param verbose: Whether to display the image after it is rendered.
    :param show_panoptic: When set to True, the lidar data is colored with the panoptic labels. When set
        to False, the colors of the lidar data represent the distance from the center of the ego vehicle.
        If show_lidarseg is True, show_panoptic will be set to False.
    """
    # lidiar_render(sample_token, pred_data, out_path=out_path)
    sample = nusc.get('sample', sample_token)
    # sample = data['results'][sample_token_list[0]][0]
    cams = [
        'CAM_FRONT_LEFT',
        'CAM_FRONT',
        'CAM_FRONT_RIGHT',
        'CAM_BACK_LEFT',
        'CAM_BACK',
        'CAM_BACK_RIGHT',
        ]

    for ind, cam in enumerate(cams):
          # 获取样本数据对应的相机数据
          sample_data_token = sample['data'][cam]

          # 获取样本数据记录
          sd_record = nusc.get('sample_data', sample_data_token)
          sensor_modality = sd_record['sensor_modality']  # 获取传感器类型

          # 仅处理相机数据
          if sensor_modality == 'camera':
              # 创建一个预测框列表,只有预测分数大于 0.5 的记录才会被包含在内
              boxes = [Box(record['translation'], record['size'], Quaternion(record['rotation']),
                           name=record['detection_name'], token='predicted') for record in
                       pred_data['results'][sample_token] if record['detection_score'] > 0.5]

              # 获取预测数据,包括图像路径、预测框和相机内参
              data_path, boxes_pred, camera_intrinsic = get_predicted_data(sample_data_token,
                                                                           box_vis_level=box_vis_level, pred_anns=boxes)
              
              _, boxes_gt, _ = nusc.get_sample_data(sample_data_token, box_vis_level=box_vis_level)

              # 打开图像数据
              image = Image.open(data_path)
              # 获取图像大小
              image_width, image_height = image.size
              
              # 获取二维边界框
              pred_boxes_2d = project_boxes_to_2d(boxes_pred, camera_intrinsic, (image_width, image_height))
              gt_boxes_2d = project_boxes_to_2d(boxes_gt, camera_intrinsic, (image_width, image_height)) 
 
              # 找出未匹配的 gt 边界框
              unmatched_gt_boxes = find_unmatched_gt(gt_boxes_2d, pred_boxes_2d, iou_threshold=1)
              # 找出已匹配的 gt 边界框
              matched_gt_boxes = find_matched_gt(gt_boxes_2d, pred_boxes_2d, iou_threshold=1)

              # 计算未匹配边界框的长宽及面积数据
              data = []
              for box in unmatched_gt_boxes:
                row_name = f"{sample_data_token}_{cam}"
                width, height, area = calculate_box_data(box)
                area_ratio = calculate_area_ratio(box,image.size)
                data.append({
                    "token":row_name,
                    'Width': width,
                    'Height': height,
                    'Area': area,
                    'Area_ratio': area_ratio
                    
                })
              
              # 将数据保存到 Excel 文件
              excel_output_path = os.path.join(save_dir, 'unmatched_gt_boxes.xlsx')
              save_to_excel(data,excel_output_path)

              # 绘制gt_2d_box
              plt.figure()
              plt.imshow(image)
              ax = plt.gca()

              for box, corners in zip(boxes_gt, unmatched_gt_boxes):
                color = np.array(get_color(box.name)) / 255.0
                for i in range(4):
                    p1 = corners[i]
                    p2 = corners[(i + 1) % 4]
                    ax.plot([p1[0], p2[0]], [p1[1], p2[1]], color=color, linewidth=2)                       

              # 限制可见范围
              plt.xlim(0, image.size[0])
              plt.ylim(image.size[1], 0)

              # 隐藏坐标轴
              plt.axis('off')
              plt.title('GT: {} {labels_type}'.format(
                sd_record['channel'], labels_type='(groundTruth)' if lidarseg_preds_bin_path else ''))
              
              # 保存或显示图像
              if out_path:
                gt_out_dir =os.path.join(save_dir,'unmatched_2dbox', out_path)
                os.makedirs(gt_out_dir, exist_ok=True)  # 确保输出目录存在
                gt_out_path = os.path.join(gt_out_dir, f'{cam}.png')
                plt.savefig(gt_out_path, bbox_inches='tight', pad_inches=0, dpi=200)

              # 保存裁剪后的图像
              output_directory = os.path.join(gt_out_dir, cam)
              crop_and_save_boxes(data_path, unmatched_gt_boxes, output_directory)

              plt.close()  # 关闭图形

              # 打开图像数据
              img = Image.open(data_path)
              # 获取图像大小
              image_width, image_height = img.size
              # 计算已匹配边界框的长宽及面积数据
              data = []
              for box in matched_gt_boxes:
                row_name = f"{sample_data_token}_{cam}"
                width, height, area = calculate_box_data(box)
                area_ratio = calculate_area_ratio(box,img.size)
                data.append({
                    "token":row_name,
                    'Width': width,
                    'Height': height,
                    'Area': area,
                    'Area_ratio': area_ratio
                })
              
              # 将数据保存到 Excel 文件
              excel_output_path = os.path.join(save_dir, 'matched_gt_boxes.xlsx')
              save_to_excel(data,excel_output_path)

              # 绘制gt_2d_box
              plt.figure()
              plt.imshow(img)
              ax = plt.gca()

              for box, corners in zip(boxes_gt, matched_gt_boxes):
                color = np.array(get_color(box.name)) / 255.0
                for i in range(4):
                    p1 = corners[i]
                    p2 = corners[(i + 1) % 4]
                    ax.plot([p1[0], p2[0]], [p1[1], p2[1]], color=color, linewidth=2)                       

              # 限制可见范围
              plt.xlim(0, img.size[0])
              plt.ylim(img.size[1], 0)

              # 隐藏坐标轴
              plt.axis('off')
              plt.title('GT: {} {labels_type}'.format(
                sd_record['channel'], labels_type='(groundTruth)' if lidarseg_preds_bin_path else ''))
              
              # 保存或显示图像
              if out_path:
                gt_out_dir = os.path.join(save_dir,'matched_2dbox', out_path)
                os.makedirs(gt_out_dir, exist_ok=True)  # 确保输出目录存在
                gt_out_path = os.path.join(gt_out_dir, f'{cam}.png')
                plt.savefig(gt_out_path, bbox_inches='tight', pad_inches=0, dpi=200)

              # 保存裁剪后的图像
              output_directory = os.path.join(gt_out_dir,cam)
              crop_and_save_boxes(data_path, matched_gt_boxes, output_directory)
              plt.close()  # 关闭图形
              

          else:
              raise ValueError("Error: Unknown sensor modality!")  # 未知传感器类型错误   
              


if __name__ == '__main__':

    nusc = NuScenes(version='v1.0-mini', dataroot='./data/nuscenes', verbose=False)

    bevformer_results = mmcv.load('val/work_dirs/bevformer_tiny/Wed_Jun_26_21_30_50_2024/pts_bbox/results_nusc.json')
    save_dir = "result"
    if not os.path.exists(save_dir):
        os.mkdir(save_dir)

    sample_token_list = list(bevformer_results['results'].keys())

    for id in range(0,10):
        render_sample_data(sample_token_list[id], pred_data=bevformer_results, out_path=sample_token_list[id])



# if verbose:
#     plt.show()

可视化结果:

 

  • 3
    点赞
  • 3
    收藏
    觉得还不错? 一键收藏
  • 0
    评论
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值