当前位置: X-MOL 学术Pattern Recogn. › 论文详情
Our official English website, www.x-mol.net, welcomes your feedback! (Note: you will need to create a separate account there.)
CAGNet: Content-Aware Guidance for Salient Object Detection
Pattern Recognition ( IF 7.5 ) Pub Date : 2020-07-01 , DOI: 10.1016/j.patcog.2020.107303
Sina Mohammadi , Mehrdad Noori , Ali Bahri , Sina Ghofrani Majelan , Mohammad Havaei

Abstract Beneficial from Fully Convolutional Neural Networks (FCNs), saliency detection methods have achieved promising results. However, it is still challenging to learn effective features for detecting salient objects in complicated scenarios, in which i) non-salient regions may have “salient-like” appearance; ii) the salient objects may have different-looking regions. To handle these complex scenarios, we propose a Feature Guide Network which exploits the nature of low-level and high-level features to i) make foreground and background regions more distinct and suppress the non-salient regions which have “salient-like” appearance; ii) assign foreground label to different-looking salient regions. Furthermore, we utilize a Multi-scale Feature Extraction Module (MFEM) for each level of abstraction to obtain multi-scale contextual information. Finally, we design a loss function which outperforms the widely used Cross-entropy loss. By adopting four different pre-trained models as the backbone, we prove that our method is very general with respect to the choice of the backbone model. Experiments on six challenging datasets demonstrate that our method achieves the state-of-the-art performance in terms of different evaluation metrics. Additionally, our approach contains fewer parameters than the existing ones, does not need any post-processing, and runs fast at a real-time speed of 28 FPS when processing a 480 × 480 image.

中文翻译:

CAGNet:显着对象检测的内容感知指南

摘要 得益于全卷积神经网络 (FCN),显着性检测方法取得了可喜的成果。然而,在复杂场景中学习用于检测显着对象的有效特征仍然具有挑战性,其中 i) 非显着区域可能具有“类似显着”的外观;ii) 显着物体可能有不同的区域。为了处理这些复杂的场景,我们提出了一个特征引导网络,它利用低级和高级特征的性质来 i) 使前景和背景区域更加清晰,并抑制具有“类似显着”外观的非显着区域; ii) 将前景标签分配给不同外观的显着区域。此外,我们对每个抽象级别使用多尺度特征提取模块(MFEM)来获取多尺度上下文信息。最后,我们设计了一个优于广泛使用的交叉熵损失的损失函数。通过采用四种不同的预训练模型作为主干,我们证明了我们的方法在主干模型的选择方面非常通用。在六个具有挑战性的数据集上的实验表明,我们的方法在不同的评估指标方面达到了最先进的性能。此外,我们的方法包含比现有方法更少的参数,不需要任何后处理,并且在处理 480 × 480 图像时以 28 FPS 的实时速度快速运行。在六个具有挑战性的数据集上的实验表明,我们的方法在不同的评估指标方面达到了最先进的性能。此外,我们的方法包含比现有方法更少的参数,不需要任何后处理,并且在处理 480 × 480 图像时以 28 FPS 的实时速度快速运行。在六个具有挑战性的数据集上的实验表明,我们的方法在不同的评估指标方面达到了最先进的性能。此外,我们的方法包含比现有方法更少的参数,不需要任何后处理,并且在处理 480 × 480 图像时以 28 FPS 的实时速度快速运行。
更新日期:2020-07-01
down
wechat
bug