当前位置: X-MOL 学术IEEE Trans. Image Process. › 论文详情
Our official English website, www.x-mol.net, welcomes your feedback! (Note: you will need to create a separate account there.)
Deep Local Feature Descriptor Learning With Dual Hard Batch Construction
IEEE Transactions on Image Processing ( IF 10.8 ) Pub Date : 2020-10-13 , DOI: 10.1109/tip.2020.3029424
Song Wang , Xin Guo , Yun Tie , Lin Qi , Ling Guan

Local feature descriptor learning aims to represent distinctive images or patches with the same local features, where their representation is invariant under different types of deformation. Recent studies have demonstrated that descriptor learning based on Convolutional Neural Network (CNN) is able to improve the matching performance significantly. However, they tend to ignore the importance of sample selection during the training process, leading to unstable quality of descriptors and learning efficiency. In this paper, a dual hard batch construction method is proposed to sample the hard matching and non-matching examples for training, improving the performance of the descriptor learning on different tasks. To construct the dual hard training batches, the matching examples with the minimum similarity are selected as the hard positive pairs. For each positive pair, the most similar non-matching example is then sampled from the generated hard positive pairs in the same batch as the corresponding negative. By sampling the hard positive pairs and the corresponding hard negatives, the hard batches are produced to force the CNN model to learn the descriptors with more efforts. In addition, based on the above dual hard batch construction, an $\ell _{2}^{2}$ triplet loss function is built for optimizing the training model. Specifically, we analyze the superiority of the $\ell _{2}^{2}$ loss function when dealing with hard examples, and also demonstrate it in the experiments. With the benefits of the proposed sampling strategy and the $\ell _{2}^{2}$ triplet loss function, our method achieves better performance compared to state-of-the-art on the reference benchmarks for different matching tasks.

中文翻译:


通过双硬批量构建进行深度局部特征描述符学习



局部特征描述符学习旨在表示具有相同局部特征的独特图像或块,其中它们的表示在不同类型的变形下保持不变。最近的研究表明,基于卷积神经网络(CNN)的描述符学习能够显着提高匹配性能。然而,他们往往忽视了训练过程中样本选择的重要性,导致描述符质量和学习效率不稳定。本文提出了一种双硬批量构建方法,对硬匹配和非匹配示例进行采样进行训练,提高了描述符学习在不同任务上的性能。为了构建双硬训练批次,选择具有最小相似度的匹配示例作为硬正例对。对于每个正例对,然后从与相应负例同一批次中生成的硬正例对中采样最相似的非匹配示例。通过对硬正例对和相应的硬负例进行采样,生成硬批次,迫使 CNN 模型更加努力地学习描述符。另外,基于上述双硬批量构建, $\ell_{2}^{2}$三元组损失函数是为了优化训练模型而建立的。具体来说,我们分析一下其优越性$\ell_{2}^{2}$处理困难示例时的损失函数,并在实验中进行了演示。 凭借所提出的抽样策略和$\ell_{2}^{2}$通过三元组损失函数,我们的方法在不同匹配任务的参考基准上与最先进的方法相比取得了更好的性能。
更新日期:2020-10-20
down
wechat
bug