当前位置: X-MOL 学术Vis. Comput. › 论文详情
Our official English website, www.x-mol.net, welcomes your feedback! (Note: you will need to create a separate account there.)
Modality-transfer generative adversarial network and dual-level unified latent representation for visible thermal Person re-identification
The Visual Computer ( IF 3.5 ) Pub Date : 2020-11-24 , DOI: 10.1007/s00371-020-02015-z
Xing Fan , Wei Jiang , Hao Luo , Weijie Mao

Visible thermal person re-identification, also known as RGB-infrared person re-identification, is an emerging cross-modality searching problem that identifies the same person from different modalities. To solve this problem, it is necessary to know what a person looks like in different modalities. Images of the same person at the same time from the same camera view in both modalities should be captured, so that similarities and differences could be discovered. However, existing datasets do not completely satisfy those requirements. Thus, a modality-transfer generative adversarial network is proposed to generate a cross-modality counterpart for a source image in the target modality, obtaining paired images for the same person. Given that query images are from one modality and gallery images are from another modality, it is necessary to produce a unified representation for both modalities so cross-modality matching could be performed. In this study, a novel dual-level unified latent representation is proposed for visible thermal person re-identification task, including an image-level patch fusion strategy and a feature-level hierarchical granularity triplet loss, producing a more general and robust unified feature embedding. Extensive experiments on both the SYSU-MM01 dataset (with visible and near-infrared images) and the RegDB dataset (with visible and far-infrared images) demonstrate the efficiency and generality of the proposed method, which achieves state-of-the-art performance. The code will be publicly released.

中文翻译:

用于可见热人重新识别的模态转移生成对抗网络和双层统一潜在表示

可见热人重识别,也称为 RGB 红外人重识别,是一种新兴的跨模态搜索问题,可从不同模态中识别同一个人。为了解决这个问题,有必要知道一个人在不同的模态下是什么样子的。应该在两种模式下从同一相机视图同时捕获同一个人的图像,以便发现异同。然而,现有的数据集并不能完全满足这些要求。因此,提出了模态转移生成对抗网络来为目标模态中的源图像生成跨模态对应物,从而获得同一个人的配对图像。鉴于查询图像来自一种模态,图库图像来自另一种模态,有必要为两种模态生成统一的表示,以便可以执行跨模态匹配。在这项研究中,为可见热人重新识别任务提出了一种新颖的双级统一潜在表示,包括图像级补丁融合策略和特征级分层粒度三元组损失,产生更通用和鲁棒的统一特征嵌入. 在 SYSU-MM01 数据集(具有可见光和近红外图像)和 RegDB 数据集(具有可见光和远红外图像)上的大量实验证明了所提出方法的效率和通用性,达到了最先进的水平表现。该代码将公开发布。针对可见热人重识别任务提出了一种新颖的双级统一潜在表示,包括图像级补丁融合策略和特征级分层粒度三元组损失,产生更通用和更强大的统一特征嵌入。在 SYSU-MM01 数据集(具有可见光和近红外图像)和 RegDB 数据集(具有可见光和远红外图像)上的大量实验证明了所提出方法的效率和通用性,达到了最先进的水平表现。该代码将公开发布。针对可见热人重识别任务提出了一种新颖的双级统一潜在表示,包括图像级补丁融合策略和特征级分层粒度三元组损失,产生更通用和更强大的统一特征嵌入。在 SYSU-MM01 数据集(具有可见光和近红外图像)和 RegDB 数据集(具有可见光和远红外图像)上的大量实验证明了所提出方法的效率和通用性,达到了最先进的水平表现。该代码将公开发布。在 SYSU-MM01 数据集(具有可见光和近红外图像)和 RegDB 数据集(具有可见光和远红外图像)上的大量实验证明了所提出方法的效率和通用性,达到了最先进的水平表现。该代码将公开发布。在 SYSU-MM01 数据集(具有可见光和近红外图像)和 RegDB 数据集(具有可见光和远红外图像)上的大量实验证明了所提出方法的效率和通用性,达到了最先进的水平表现。该代码将公开发布。
更新日期:2020-11-24
down
wechat
bug