无人机视觉 drone

http://www.aiskyeye.com/

2018年已经办过一年了。2019年在ICCV上办。

We encourage the participants to use the provided training data for each task, but also allow them to use additional training data. The use of additional training data must be indicated in the "method description" when uploading results to the server.

We emphasize that any form of annotation or use of the VisDrone testing sets for either supervised or unsupervised training is strictly forbidden. The participants are required to explicitly specify any an all external data used for training in the "method description" in submission. In addition, the participants are NOT allowed to train a model in one task using the training or validation sets in other tasks.

we have divided the test set into two splits, including test-challenge and test-dev. Test-dev一天交3次,test-challenge一共交3次。

不明白为什么download里test-dev是不能下载的。。。

 

Vision Meets Drones: A Challenge

Abstract

our benchmark has more than 2:5 million annotated instances in 179; 264 images/video frames.

Introduction

现状:缺少大的数据集。

Altogether we carefully annotated more than 2:5 million bounding boxes of object instances from these categories. Moreover, some important attributes including visibility of scenes, object category and occlusion, are provided for better data usage.

感觉还是挺大的,也没什么别的数据集可以利用了。

3.2 Task 1: Object Detection in Images

The VisDrone2018 provides a dataset of 10; 209 images for this task, with 6; 471 images used for training, 548 for validation and 3; 190 for testing.

For truncation ratio, it is used to indicate the degree of object parts appears outside a frame. If an object is not fully captured within a frame, we annotate the bounding box across the frame boundary and estimate the truncation ratio based on the region outside the image. It is worth mentioning that a target is skipped during evaluation if its truncation ratio is larger than 50%.

task1 遮挡情况

Three degrees of occlusions: no occlusion (occlusion ratio 0%), partial occlusion (occlusion ratio 1% 50%), and heavy occlusion (occlusion ratio > 50%).

2018task1winner

选择retina net,移除后面两层,只用P3,4,5

数据分布

 

Cascade R-CNN: Delving into High Quality Object Detection

Single-Shot Bidirectional Pyramid Networks for High-Quality Object Detection

数据观察

task1

bbox观察

数据分布
bbox尺度分布
 <p2p2p3p4p5p6>p6
train48.3%30.6%16.2%4.4%0.5%0.01%0%
val55.3%30.8%11.4%2.4%0.1%0.007%0%
train_val49.0%30.6%15.7%4.1%0.5%0.01%0%

 

bbox尺度*1.5分布
 <p2p2p3p4p5p6>p6
train28.2%34.6%24.5%10.5%2.0%0.1%0.002%
val33.0%37.6%21.8%6.9%0.7%0.03%0.005%
train_val28.7%34.9%24.2%10.2%1.9%0.1%0.002%
bbox尺度*3分布
 <p2p2p3p4p5p6>p6
train6.2%22.0%34.6%24.5%10.6%2.0%0.1%
val7.1%26.0%37.6%21.8%6.9%0.7%0.03%
train_val6.3%22.4%34.9%24.2%10.2%1.9%0.1%


类别分布

类别数量分布
 0 ignored1 pedestrain2 people3 bicycle4 car5 van6 truck7 tricycly8 awning-tricycle9 bus10 motor11 others
train2.49%22.44%7.65%2.96%40.97%7.06%3.64%1.36%0.92%1.68%8.39%0.43%
val3.43%22.02%12.76%3.24%35.01%4.92%1.87%2.60%1.32%0.62%12.16%0.08%
trian_val2.59%22.40%8.17%2.99%40.37%6.84%3.46%1.49%0.96%1.57%8.77%0.40%

 

类别面积分布
 0 ignored1 pedestrain2 people3 bicycle4 car5 van6 truck7 tricycly8 awning-tricycle9 bus10 motor11 others
train3.89%6.42%1.82%1.45%54.11%11.66%9.46%1.39%1.07%4.69%3.35%0.69%
val8.23%7.44%3.67%1.42%54.88%7.68%5.80%2.38%1.29%1.82%5.23%0.15%
trian_val4.18%6.49%1.94%1.45%54.16%11.39%9.22%1.46%1.09%4.50%3.47%0.65%


图片尺寸分布

 480×360960×5401344×7561360×7651389×10421398×10481400×7881400×10501916×10781920×10802000×1500all
train12501743130129924985373397726462
val0121

0

40800000190548
test0640150007124911273601580

train 平均图像大小:1575738.0529247911

 

b0说task2里同一个物体的类别会变。

Pyramid Scene Parsing Network

  • 0
    点赞
  • 0
    收藏
    觉得还不错? 一键收藏
  • 0
    评论
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值