当前位置: X-MOL 学术IEEE Trans. Image Process. › 论文详情
Our official English website, www.x-mol.net, welcomes your feedback! (Note: you will need to create a separate account there.)
Weakly-Supervised Salient Object Detection on Light Fields
IEEE Transactions on Image Processing ( IF 10.6 ) Pub Date : 2022-09-23 , DOI: 10.1109/tip.2022.3207605
Zijian Liang 1 , Pengjie Wang 1 , Ke Xu 2 , Pingping Zhang 3 , Rynson W.H. Lau 2
Affiliation  

Most existing salient object detection (SOD) methods are designed for RGB images and do not take advantage of the abundant information provided by light fields. Hence, they may fail to detect salient objects of complex structures and delineate their boundaries. Although some methods have explored multi-view information of light field images for saliency detection, they require tedious pixel-level manual annotations of ground truths. In this paper, we propose a novel weakly-supervised learning framework for salient object detection on light field images based on bounding box annotations. Our method has two major novelties. First, given an input light field image and a bounding-box annotation indicating the salient object, we propose a ground truth label hallucination method to generate a pixel-level pseudo saliency map, to avoid heavy cost of pixel-level annotations. This method generates high quality pseudo ground truth saliency maps to help supervise the training, by exploiting information obtained from the light field (including depths and RGB images). Second, to exploit the multi-view nature of the light field data in learning, we propose a fusion attention module to calibrate the spatial and channel-wise light field representations. It learns to focus on informative features and suppress redundant information from the multi-view inputs. Based on these two novelties, we are able to train a new salient object detector with two branches in a weakly-supervised manner. While the RGB branch focuses on modeling the color contrast in the all-in-focus image for locating the salient objects, the Focal branch exploits the depth and the background spatial redundancy of focal slices for eliminating background distractions. Extensive experiments show that our method outperforms existing weakly-supervised methods and most fully supervised methods.

中文翻译:

光场上的弱监督显着目标检测

大多数现有的显着目标检测(SOD)方法都是为 RGB 图像设计的,并没有利用光场提供的丰富信息。因此,它们可能无法检测到复杂结构的显着对象并划定它们的边界。尽管一些方法已经探索了用于显着性检测的光场图像的多视图信息,但它们需要对基本事实进行繁琐的像素级手动注释。在本文中,我们提出了一种新的弱监督学习框架,用于基于边界框注释的光场图像上的显着目标检测。我们的方法有两个主要的新颖之处。首先,给定输入光场图像和指示显着对象的边界框注释,我们提出了一种地面实况标签幻觉方法来生成像素级伪显着图,以避免像素级注释的高昂成本。该方法通过利用从光场(包括深度和 RGB 图像)获得的信息,生成高质量的伪地面真实显着性图来帮助监督训练。其次,为了在学习中利用光场数据的多视图特性,我们提出了一个融合注意力模块来校准空间和通道方式的光场表示。它学会了关注信息特征并抑制来自多视图输入的冗余信息。基于这两个新颖性,我们能够以弱监督的方式训练具有两个分支的新显着对象检测器。虽然 RGB 分支专注于对全焦点图像中的颜色对比度进行建模以定位显着对象,焦点分支利用焦点切片的深度和背景空间冗余来消除背景干扰。大量实验表明,我们的方法优于现有的弱监督方法和大多数完全监督的方法。
更新日期:2022-09-23
down
wechat
bug