当前位置: X-MOL 学术ACM Trans. Multimed. Comput. Commun. Appl. › 论文详情
Our official English website, www.x-mol.net, welcomes your feedback! (Note: you will need to create a separate account there.)
Region-Level Visual Consistency Verification for Large-Scale Partial-Duplicate Image Search
ACM Transactions on Multimedia Computing, Communications, and Applications ( IF 5.1 ) Pub Date : 2020-05-25 , DOI: 10.1145/3383582
Zhili Zhou 1 , Q. M. Jonathan Wu 2 , Yimin Yang 3 , Xingming Sun 1
Affiliation  

Most recent large-scale image search approaches build on a bag-of-visual-words model, in which local features are quantized and then efficiently matched between images. However, the limited discriminability of local features and the BOW quantization errors cause a lot of mismatches between images, which limit search accuracy. To improve the accuracy, geometric verification is popularly adopted to identify geometrically consistent local matches for image search, but it is hard to directly use these matches to distinguish partial-duplicate images from non-partial-duplicate images. To address this issue, instead of simply identifying geometrically consistent matches, we propose a region-level visual consistency verification scheme to confirm whether there are visually consistent region (VCR) pairs between images for partial-duplicate search. Specifically, after the local feature matching, the potential VCRs are constructed via mapping the regions segmented from candidate images to a query image by utilizing the properties of the matched local features. Then, the compact gradient descriptor and convolutional neural network descriptor are extracted and matched between the potential VCRs to verify their visual consistency to determine whether they are VCRs. Moreover, two fast pruning algorithms are proposed to further improve efficiency. Extensive experiments demonstrate the proposed approach achieves higher accuracy than the state of the art and provide comparable efficiency for large-scale partial-duplicate search tasks.

中文翻译:

大规模部分重复图像搜索的区域级视觉一致性验证

最近的大规模图像搜索方法建立在视觉词袋模型上,其中局部特征被量化,然后在图像之间进行有效匹配。然而,局部特征的有限可辨别性和 BOW 量化误差导致图像之间存在大量不匹配,从而限制了搜索精度。为了提高准确性,几何验证被广泛用于识别几何一致的局部匹配进行图像搜索,但很难直接使用这些匹配来区分部分重复图像和非部分重复图像。为了解决这个问题,我们提出了一种区域级视觉一致性验证方案,而不是简单地识别几何上一致的匹配,以确认图像之间是否存在视觉一致区域(VCR)对以进行部分重复搜索。具体来说,在局部特征匹配之后,通过利用匹配的局部特征的属性将从候选图像分割的区域映射到查询图像来构建潜在的VCR。然后,提取紧凑梯度描述符和卷积神经网络描述符,并在潜在 VCR 之间进行匹配,以验证它们的视觉一致性,以确定它们是否是 VCR。此外,提出了两种快速剪枝算法以进一步提高效率。大量实验表明,所提出的方法比现有技术实现了更高的准确性,并为大规模的部分重复搜索任务提供了相当的效率。利用匹配的局部特征的属性,通过将从候选图像分割的区域映射到查询图像来构建潜在的 VCR。然后,提取紧凑梯度描述符和卷积神经网络描述符,并在潜在 VCR 之间进行匹配,以验证它们的视觉一致性,以确定它们是否是 VCR。此外,提出了两种快速剪枝算法以进一步提高效率。大量实验表明,所提出的方法比现有技术实现了更高的准确性,并为大规模的部分重复搜索任务提供了相当的效率。利用匹配的局部特征的属性,通过将从候选图像分割的区域映射到查询图像来构建潜在的 VCR。然后,提取紧凑梯度描述符和卷积神经网络描述符,并在潜在 VCR 之间进行匹配,以验证它们的视觉一致性,以确定它们是否是 VCR。此外,提出了两种快速剪枝算法以进一步提高效率。大量实验表明,所提出的方法比现有技术实现了更高的准确性,并为大规模的部分重复搜索任务提供了相当的效率。压缩梯度描述符和卷积神经网络描述符被提取并在潜在VCR之间进行匹配,以验证它们的视觉一致性以确定它们是否是VCR。此外,提出了两种快速剪枝算法以进一步提高效率。大量实验表明,所提出的方法比现有技术实现了更高的准确性,并为大规模的部分重复搜索任务提供了相当的效率。压缩梯度描述符和卷积神经网络描述符被提取并在潜在VCR之间进行匹配,以验证它们的视觉一致性以确定它们是否是VCR。此外,提出了两种快速剪枝算法以进一步提高效率。大量实验表明,所提出的方法比现有技术实现了更高的准确性,并为大规模的部分重复搜索任务提供了相当的效率。
更新日期:2020-05-25
down
wechat
bug