道路行人数据集汇总

原链接:https://blog.csdn.net/weixin_38907560/article/details/81266462

1、Caltech Pedestrian Detection Benchmark

http://www.vision.caltech.edu/Image_Datasets/CaltechPedestrians/ 加州理工学院行人数据集包含大约10个小时的640x480 30Hz视频,这些视频是通过在城市环境中正常行驶的车辆拍摄的。大约250,000帧(137分钟左右),共计350,000个边界框和2300个独特的行人被注释。注释包括边界框和详细的遮挡标签之间的时间对应关系。

2、Penn-Fudan Database for Pedestrian Detection and Segmentation

https://www.cis.upenn.edu/~jshi/ped_html/ 这是一个包含图像的图像数据库。图像取自校园和城市街道周围的场景。每幅图像至少有一个行人。这个数据库的像素为[180,390]。所有标记的行人都是站立的。共有170幅图片,其中345人被标记为行人。

3、Daimler Pedestrain Benchmark Data Sets (有8个数据集)

http://www.gavrila.net/Datasets/Daimler_Pedestrian_Benchmark_D/daimler_pedestrian_benchmark_d.html 该网址包含了戴姆勒许多数据集;戴姆勒行人分割基准数据集;戴姆勒行人预测基准数据集;戴姆勒立体行人检测基准数据集;戴姆勒单行道行人检测数据集;等等

4、GM-ATCI Rear-view pedestrians dataset

https://sites.google.com/site/rearviewpeds1/ 从鱼眼镜头摄像头拍摄的后视行人数据集。这个数据集包含了250个视频片段,总时长为76分钟,超过200 K标记的行人边界框。
————————————————
版权声明:本文为CSDN博主「律者无疆」的原创文章,遵循 CC 4.0 BY-SA 版权协议,转载请附上原文出处链接及本声明。
原文链接:https://blog.csdn.net/weixin_38907560/article/details/81266462

  • 0
    点赞
  • 11
    收藏
    觉得还不错? 一键收藏
  • 0
    评论
Human parsing has been extensively studied recently (Yamaguchi et al. 2012; Xia et al. 2017) due to its wide applications in many important scenarios. Mainstream fashion parsing models (i.e., parsers) focus on parsing the high-resolution and clean images. However, directly applying the parsers trained on benchmarks of high-quality samples to a particular application scenario in the wild, e.g., a canteen, airport or workplace, often gives non-satisfactory performance due to domain shift. In this paper, we explore a new and challenging cross-domain human parsing problem: taking the benchmark dataset with extensive pixel-wise labeling as the source domain, how to obtain a satisfactory parser on a new target domain without requiring any additional manual labeling? To this end, we propose a novel and efficient crossdomain human parsing model to bridge the cross-domain differences in terms of visual appearance and environment conditions and fully exploit commonalities across domains. Our proposed model explicitly learns a feature compensation network, which is specialized for mitigating the cross-domain differences. A discriminative feature adversarial network is introduced to supervise the feature compensation to effectively reduces the discrepancy between feature distributions of two domains. Besides, our proposed model also introduces a structured label adversarial network to guide the parsing results of the target domain to follow the high-order relationships of the structured labels shared across domains. The proposed framework is end-to-end trainable, practical and scalable in real applications. Extensive experiments are conducted where LIP dataset is the source domain and 4 different datasets including surveillance videos, movies and runway shows without any annotations, are evaluated as target domains. The results consistently confirm data efficiency and performance advantages of the proposed method for the challenging cross-domain human parsing problem. Abstract—This paper presents a robust Joint Discriminative appearance model based Tracking method using online random forests and mid-level feature (superpixels). To achieve superpixel- wise discriminative ability, we propose a joint appearance model that consists of two random forest based models, i.e., the Background-Target discriminative Model (BTM) and Distractor- Target discriminative Model (DTM). More specifically, the BTM effectively learns discriminative information between the target object and background. In contrast, the DTM is used to suppress distracting superpixels which significantly improves the tracker’s robustness and alleviates the drifting problem. A novel online random forest regression algorithm is proposed to build the two models. The BTM and DTM are linearly combined into a joint model to compute a confidence map. Tracking results are estimated using the confidence map, where the position and scale of the target are estimated orderly. Furthermore, we design a model updating strategy to adapt the appearance changes over time by discarding degraded trees of the BTM and DTM and initializing new trees as replacements. We test the proposed tracking method on two large tracking benchmarks, the CVPR2013 tracking benchmark and VOT2014 tracking challenge. Experimental results show that the tracker runs at real-time speed and achieves favorable tracking performance compared with the state-of-the-art methods. The results also sug- gest that the DTM improves tracking performance significantly and plays an important role in robust tracking.
YOLox行人检测数据集是一个用于训练和评估行人检测算法的数据集。它是由一组带有标注框的图像组成,每个标注框都表示图像中一个行人的位置和边界。此数据集的目的是为了帮助研究人员和开发者开发更准确、高效的行人检测算法。 YOLox行人检测数据集具有以下特点。首先,它的图像数量较大,能够提供充足的训练数据,从而帮助算法更好地理解行人的外观和特征。其次,标注框的质量较高,具有准确的位置和边界信息,这有助于算法准确地定位行人。此外,数据集中的图像场景也多样化,包括室内外、不同天气和光照条件下的行人图像,从而使算法具备更好的泛化能力。 通过使用YOLox行人检测数据集,研究人员和开发者可以训练和调整行人检测算法,从而实现在图像或视频中检测和跟踪行人的应用。例如,在监控系统中,该算法可以实时检测异常行为或危险情况,提升安全性。此外,行人检测算法也可以应用于自动驾驶系统中,帮助车辆更好地感知和识别道路上的行人,提高行车安全性。 综上所述,YOLox行人检测数据集是一个用于训练和评估行人检测算法的重要数据资源,它具备大量的图像样本、准确的标注和多样的场景,有助于提升行人检测算法的准确性和鲁棒性,为各种应用场景提供更好的行人检测能力。

“相关推荐”对你有帮助么?

  • 非常没帮助
  • 没帮助
  • 一般
  • 有帮助
  • 非常有帮助
提交
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值