当前位置: X-MOL 学术Connect. Sci. › 论文详情
Our official English website, www.x-mol.net, welcomes your feedback! (Note: you will need to create a separate account there.)
Aggregating multi-scale contextual features from multiple stages for semantic image segmentation
Connection Science ( IF 3.2 ) Pub Date : 2021-02-10 , DOI: 10.1080/09540091.2020.1862059
Dingchao Jiang, Hua Qu, Jihong Zhao, Jianlong Zhao, Meng-Yen Hsieh

Semantic segmentation plays a vital role in image understanding. Recent studies have attempted to achieve precise pixel-level classification by using deep networks that provide hierarchical features. These methods are trying to effectively utilise multi-level features that are extracted from the data and precisely reconstruct some characteristics of objects that are lost in producing high-level features. In this paper, we propose a multi-scale context U-net (MSCU-net) for semantic image segmentation. This network uses a multi-scale context block (MSCB) to aggregate multi-level features and employs the CRF layer to explicitly model the dependencies among pixels. This network significantly outperforms other state-of-the-art methods on both the PASCAL VOC 2012 and Cityscapes datasets.



中文翻译:

从多个阶段聚合多尺度上下文特征以进行语义图像分割

语义分割在图像理解中起着至关重要的作用。最近的研究试图通过使用提供分层特征的深度网络来实现精确的像素级分类。这些方法试图有效地利用从数据中提取的多级特征,并精确重建在生成高级特征时丢失的对象的某些特征。在本文中,我们提出了一种用于语义图像分割的多尺度上下文 U-net(MSCU-net)。该网络使用多尺度上下文块 (MSCB) 来聚合多级特征,并使用 CRF 层对像素之间的依赖关系进行显式建模。该网络在 PASCAL VOC 2012 和 Cityscapes 数据集上的性能明显优于其他最先进的方法。

更新日期:2021-02-10
down
wechat
bug