阿里行人重识别(ReID)算法效果取得世界第一

近日,阿里巴巴机器智能技术实验室在行人重识别(Person Re-identification)算法上获得突破性成果。该技术团队通过最新算法,在不使用任何时序信息的情况下,在行人重识别主流数据库(Market1501,DukeMTMC-reid和CUHK03)上各项指标均取得第一的好成绩,刷新了业内的最好成绩。


阿里行人重识别(ReID)算法效果取得世界第一


其性能的提高主要来源于技术层面的创新:该团队通过局部信息的挖掘,致力于解决行人在识别过程中表观姿态变化剧烈,不容易对齐的问题。一方面,通过人体语义分割得到具有强语义信息的部件,并利用注意力机制在其中寻找最具有区分性的区域。另一方面,使用了基于金字塔的水平分块策略,得到行人固定区域的可辨识信息。在训练中,同时采用两种策略相结合的方式,达到行人图片的对齐,从而实现更精准的匹配识别。通过技术上的改进,该方法在三个公开数据库上的效果均优于之前最好方法,特别是mAP指标,分别提升了2%,1.87%,3.39%。


阿里行人重识别(ReID)算法效果取得世界第一


近年来,行人重识别技术在业内得到了越来越多的关注,仅CVPR2018就有将近30多篇文章专注于行人重识别问题的研究。随着行人重识别技术的日渐成熟,其巨大的应用价值和市场潜力得到了越来越多的关注。

不管是老牌的安防公司如海康威视,浙江大华,还是新晋独角兽旷世科技,商汤科技,还有传统的互联网巨头,BAT,华为等,都对行人重识别非常关注,在算法,数据和人才等各个方面上进行着布局和积累。此次阿里巴巴机器智能技术实验室在行人重识别算法的突破,为其在新零售领域相关技术方案落地奠定了稳固基础。

  • 0
    点赞
  • 10
    收藏
    觉得还不错? 一键收藏
  • 3
    评论
Human parsing has been extensively studied recently (Yamaguchi et al. 2012; Xia et al. 2017) due to its wide applications in many important scenarios. Mainstream fashion parsing models (i.e., parsers) focus on parsing the high-resolution and clean images. However, directly applying the parsers trained on benchmarks of high-quality samples to a particular application scenario in the wild, e.g., a canteen, airport or workplace, often gives non-satisfactory performance due to domain shift. In this paper, we explore a new and challenging cross-domain human parsing problem: taking the benchmark dataset with extensive pixel-wise labeling as the source domain, how to obtain a satisfactory parser on a new target domain without requiring any additional manual labeling? To this end, we propose a novel and efficient crossdomain human parsing model to bridge the cross-domain differences in terms of visual appearance and environment conditions and fully exploit commonalities across domains. Our proposed model explicitly learns a feature compensation network, which is specialized for mitigating the cross-domain differences. A discriminative feature adversarial network is introduced to supervise the feature compensation to effectively reduces the discrepancy between feature distributions of two domains. Besides, our proposed model also introduces a structured label adversarial network to guide the parsing results of the target domain to follow the high-order relationships of the structured labels shared across domains. The proposed framework is end-to-end trainable, practical and scalable in real applications. Extensive experiments are conducted where LIP dataset is the source domain and 4 different datasets including surveillance videos, movies and runway shows without any annotations, are evaluated as target domains. The results consistently confirm data efficiency and performance advantages of the proposed method for the challenging cross-domain human parsing problem. Abstract—This paper presents a robust Joint Discriminative appearance model based Tracking method using online random forests and mid-level feature (superpixels). To achieve superpixel- wise discriminative ability, we propose a joint appearance model that consists of two random forest based models, i.e., the Background-Target discriminative Model (BTM) and Distractor- Target discriminative Model (DTM). More specifically, the BTM effectively learns discriminative information between the target object and background. In contrast, the DTM is used to suppress distracting superpixels which significantly improves the tracker’s robustness and alleviates the drifting problem. A novel online random forest regression algorithm is proposed to build the two models. The BTM and DTM are linearly combined into a joint model to compute a confidence map. Tracking results are estimated using the confidence map, where the position and scale of the target are estimated orderly. Furthermore, we design a model updating strategy to adapt the appearance changes over time by discarding degraded trees of the BTM and DTM and initializing new trees as replacements. We test the proposed tracking method on two large tracking benchmarks, the CVPR2013 tracking benchmark and VOT2014 tracking challenge. Experimental results show that the tracker runs at real-time speed and achieves favorable tracking performance compared with the state-of-the-art methods. The results also sug- gest that the DTM improves tracking performance significantly and plays an important role in robust tracking.
评论 3
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值