VALSE学习(二):行人重识别研究进展回顾

VALSE2019报告  澳大利亚国立大学-郑良

一、行人重识别定义

Person Re-identification也称行人再识别,简称为ReID,是利用计算机视觉技术判断图像或者视频序列中是否存在特定行人的技术。广泛被认为是一个图像检索的子问题。给定一个监控行人图像,检索跨设备的该行人图像。

在监控视频中,由于相机分辨率和拍摄角度的缘故,通常无法得到质量非常高的人脸图片。当人脸识别失效的情况下,ReID就成为了一个非常重要的替代品技术。

数据集通常是通过人工标注或者检测算法得到的行人图片,目前与检测独立,注重识别

数据集分为训练集、验证集、Query、Gallery

在训练集上进行模型的训练,得到模型后对Query与Gallery中的图片提取特征计算相似度,对于每个Query在Gallery中找出前N

个与其相似的图片,训练、测试中人物身份不重复。

二、主要方法

特征提取:学习能够应对在不同摄像头下行人变化的特征。

度量学习 :将学习到的特征映射到新的空间使相同的人更近,不同的人更远。

三、论文回顾

行人再识别是通用物体检索的特殊情况,也是多目标多相机跟踪的核心组成部分,具有重要的研究和应用价值。行人再识别近年来取得了长足的发展,在多个数据集上实现了较高的识别精度,并在工业界得到了较多落地级别的应用。目前行人再识别的研究内容主要集中在几个方面,如网络设计、充分利用行人特性、提高模型泛化能力等。本次报告将总结行人再识别在过去一年中的进展,包括性能提升(单标准数据集和跨数据集),新任务的衍生(基于视频、跟踪、属性识别、检测与识别的结合),具有重要意义的新结论的提出等等。最后,将简短展望行人再识别领域未来的发展方向。

概述:(1)同一个域的训练集和测试集(2)无监督域自适应(3)行人搜索

行人重识别定义:在原始视频流中进行行人检测,利用camera1的单个行人作为探测集,在gallery的camera中进行查找,找出相似的结果。行人重新识别问题=行人检测+跨摄像头追踪。

part_level:利用分块的思想进行对比。

PCB main contribution:

1. 提出了均匀分块的 Part-based Convolutional Baseline(PCB),探讨了较优的块间组合方式。

2. 提出了基于 parts 的 Refined part pooling(RPP),用注意力机制来对齐 parts。

stripe partition:条纹分区,采用根据纹理分块进行卷积提取特征。

利用姿态和3D信息进行处理。

smooth manifold:有监督的光滑流形

在线实例匹配

 

 

 

多目标多相机追踪

多目标追踪

 

Human parsing has been extensively studied recently (Yamaguchi et al. 2012; Xia et al. 2017) due to its wide applications in many important scenarios. Mainstream fashion parsing models (i.e., parsers) focus on parsing the high-resolution and clean images. However, directly applying the parsers trained on benchmarks of high-quality samples to a particular application scenario in the wild, e.g., a canteen, airport or workplace, often gives non-satisfactory performance due to domain shift. In this paper, we explore a new and challenging cross-domain human parsing problem: taking the benchmark dataset with extensive pixel-wise labeling as the source domain, how to obtain a satisfactory parser on a new target domain without requiring any additional manual labeling? To this end, we propose a novel and efficient crossdomain human parsing model to bridge the cross-domain differences in terms of visual appearance and environment conditions and fully exploit commonalities across domains. Our proposed model explicitly learns a feature compensation network, which is specialized for mitigating the cross-domain differences. A discriminative feature adversarial network is introduced to supervise the feature compensation to effectively reduces the discrepancy between feature distributions of two domains. Besides, our proposed model also introduces a structured label adversarial network to guide the parsing results of the target domain to follow the high-order relationships of the structured labels shared across domains. The proposed framework is end-to-end trainable, practical and scalable in real applications. Extensive experiments are conducted where LIP dataset is the source domain and 4 different datasets including surveillance videos, movies and runway shows without any annotations, are evaluated as target domains. The results consistently confirm data efficiency and performance advantages of the proposed method for the challenging cross-domain human parsing problem. Abstract—This paper presents a robust Joint Discriminative appearance model based Tracking method using online random forests and mid-level feature (superpixels). To achieve superpixel- wise discriminative ability, we propose a joint appearance model that consists of two random forest based models, i.e., the Background-Target discriminative Model (BTM) and Distractor- Target discriminative Model (DTM). More specifically, the BTM effectively learns discriminative information between the target object and background. In contrast, the DTM is used to suppress distracting superpixels which significantly improves the tracker’s robustness and alleviates the drifting problem. A novel online random forest regression algorithm is proposed to build the two models. The BTM and DTM are linearly combined into a joint model to compute a confidence map. Tracking results are estimated using the confidence map, where the position and scale of the target are estimated orderly. Furthermore, we design a model updating strategy to adapt the appearance changes over time by discarding degraded trees of the BTM and DTM and initializing new trees as replacements. We test the proposed tracking method on two large tracking benchmarks, the CVPR2013 tracking benchmark and VOT2014 tracking challenge. Experimental results show that the tracker runs at real-time speed and achieves favorable tracking performance compared with the state-of-the-art methods. The results also sug- gest that the DTM improves tracking performance significantly and plays an important role in robust tracking.
评论 4
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值