行人重识别自制数据集划分

1、尝试:每个ID的20%的图片划分到query.txt中,其余80%划分到gallery.txt中。

2、 
 DeepReID: Deep Filter Pairing Neural Network for Person Re-Identification

ID:1360 ,训练集 1160 个,验证集 100 个,测试集 100 个。测试协议是 single-shot setting,即每个行人在 gallery 里只有一张图片。重复执行了 20 次上述随机划分。

3、Re-ranking person re-identification with k-reciprocal encoding

ID:1467 ,训练集中有 767 个行人,测试集中有 700 个行人。测试协议是multi-shot ,每个行人在 gallery 集中有多张图片。随机选取一张图片,构成 query 集,不同于旧的测试协议,新的测试协议只执行一次训练集/测试集划分。

4、Scalable person re-identification: A benchmark

行人图片采集自清华大学校园的6个摄像头(其中有5个高清摄像头和1个低分辨率摄像头),一共标注了1501个不同身份的行人,通过DPM检测器生成行人检测框,共32668张图像。训练集中包含12936张行人图像,对应751个不同身份的行人;测试集中包含19732张图像,对应750个不同身份的行人。另外,该数据集分别从6个摄像头对应的测试集中为每个行人选取一张图像构成查询图像集,包含3368张query图片,这些图片的行人检测框由人工裁剪得到,以保证检索目标的正确性。之后,为了模拟大规模检索,加入了约50万幅干扰图像,包含了失败的检测框、不在测试集当中的行人等,使该数据集更贴近真实场景。
 

5、DukeMTMC4ReID: A Large-Scale Multi-Camera Person Re-Identification Dataset

Duke大学在2017年发布的行人重识别数据集。行人图片由8个高清摄像头采集,包括1812个不同身份的行人,共36411张图片,其中至少在两个以上摄像头中出现的行人有1404人,只在一个摄像头中出现的行人有408人。训练集中包含了来自出现在多摄像头中的702个行人的16522张图片,测试集中包含出现在多摄像头中的另外702个行人和只出现在单摄像头中的408个行人的17661张图片,查询图像集包含2228张图像。query分法为取其中一个摄像头的图像作为query。

6、Person re-identification by contour sketch under moderate clothing change

包含211个行人的33 698张室内环境下的图像,每个人平均图像数量约为152张,作者没有提到具体的数据集划分方法。

7、DukeMTMC4ReID: A Large-Scale Multi-Camera Person Re-Identification Dataset

包含 4,101 个身份,共 126,441 张图像。训练集和测试集是由不同的摄像头采集的图像组成。测试集包含多种不同的场景,包括户外和室内。将数据集中的图像分为训练集和测试集。在测试集中,将一部分身份作为查询身份,另一部分作为图库身份。通常的做法是使用随机划分或者按照相机视角划分。

Human parsing has been extensively studied recently (Yamaguchi et al. 2012; Xia et al. 2017) due to its wide applications in many important scenarios. Mainstream fashion parsing models (i.e., parsers) focus on parsing the high-resolution and clean images. However, directly applying the parsers trained on benchmarks of high-quality samples to a particular application scenario in the wild, e.g., a canteen, airport or workplace, often gives non-satisfactory performance due to domain shift. In this paper, we explore a new and challenging cross-domain human parsing problem: taking the benchmark dataset with extensive pixel-wise labeling as the source domain, how to obtain a satisfactory parser on a new target domain without requiring any additional manual labeling? To this end, we propose a novel and efficient crossdomain human parsing model to bridge the cross-domain differences in terms of visual appearance and environment conditions and fully exploit commonalities across domains. Our proposed model explicitly learns a feature compensation network, which is specialized for mitigating the cross-domain differences. A discriminative feature adversarial network is introduced to supervise the feature compensation to effectively reduces the discrepancy between feature distributions of two domains. Besides, our proposed model also introduces a structured label adversarial network to guide the parsing results of the target domain to follow the high-order relationships of the structured labels shared across domains. The proposed framework is end-to-end trainable, practical and scalable in real applications. Extensive experiments are conducted where LIP dataset is the source domain and 4 different datasets including surveillance videos, movies and runway shows without any annotations, are evaluated as target domains. The results consistently confirm data efficiency and performance advantages of the proposed method for the challenging cross-domain human parsing problem. Abstract—This paper presents a robust Joint Discriminative appearance model based Tracking method using online random forests and mid-level feature (superpixels). To achieve superpixel- wise discriminative ability, we propose a joint appearance model that consists of two random forest based models, i.e., the Background-Target discriminative Model (BTM) and Distractor- Target discriminative Model (DTM). More specifically, the BTM effectively learns discriminative information between the target object and background. In contrast, the DTM is used to suppress distracting superpixels which significantly improves the tracker’s robustness and alleviates the drifting problem. A novel online random forest regression algorithm is proposed to build the two models. The BTM and DTM are linearly combined into a joint model to compute a confidence map. Tracking results are estimated using the confidence map, where the position and scale of the target are estimated orderly. Furthermore, we design a model updating strategy to adapt the appearance changes over time by discarding degraded trees of the BTM and DTM and initializing new trees as replacements. We test the proposed tracking method on two large tracking benchmarks, the CVPR2013 tracking benchmark and VOT2014 tracking challenge. Experimental results show that the tracker runs at real-time speed and achieves favorable tracking performance compared with the state-of-the-art methods. The results also sug- gest that the DTM improves tracking performance significantly and plays an important role in robust tracking.
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值