当前位置: X-MOL 学术Comput. Biol. Med. › 论文详情
Our official English website, www.x-mol.net, welcomes your feedback! (Note: you will need to create a separate account there.)
CSAP-UNet: Convolution and self-attention paralleling network for medical image segmentation with edge enhancement
Computers in Biology and Medicine ( IF 7.7 ) Pub Date : 2024-03-07 , DOI: 10.1016/j.compbiomed.2024.108265
Xiaodong Fan , Jing Zhou , Xiaoli Jiang , Meizhuo Xin , Limin Hou

Convolution operation is performed within a local window of the input image. Therefore, convolutional neural network (CNN) is skilled in obtaining local information. Meanwhile, the self-attention (SA) mechanism extracts features by calculating the correlation between tokens from all positions in the image, which has advantage in obtaining global information. Therefore, the two modules can complement each other to improve feature extraction ability. An effective fusion method is a problem worthy of further study. In this paper, we propose a CNN and SA paralleling network CSAP-UNet with U-Net as backbone. The encoder consists of two parallel branches of CNN and Transformer to extract the feature from the input image, which takes into account both the global dependencies and the local information. Because medical images come from certain frequency bands within the spectrum, their color channels are not as uniform as natural images. Meanwhile, medical segmentation pays more attention to lesion regions in the image. Attention fusion module (AFM) integrates channel attention and spatial attention in series to fuse the output features of the two branches. The medical image segmentation task is essentially to locate the boundary of the object in the image. The boundary enhancement module (BEM) is designed in the shallow layer of the proposed network to focus more specifically on pixel-level edge details. Experimental results on three public datasets validate that CSAP-UNet outperforms state-of-the-art networks, particularly on the ISIC 2017 dataset. The cross-dataset evaluation on Kvasir and CVC-ClinicDB shows that CSAP-UNet has strong generalization ability. Ablation experiments also indicate the effectiveness of the designed modules. The code for training and test is available at .

中文翻译:

CSAP-UNet:用于具有边缘增强的医学图像分割的卷积和自注意力并行网络

卷积运算在输入图像的局部窗口内执行。因此,卷积神经网络(CNN)擅长获取局部信息。同时,自注意力(SA)机制通过计算图像中所有位置的token之间的相关性来提取特征,在获取全局信息方面具有优势。因此,两个模块可以相互补充,提高特征提取能力。有效的融合方法是一个值得进一步研究的问题。在本文中,我们提出了一种以 U-Net 作为骨干的 CNN 和 SA 并行网络 CSAP-UNet。编码器由 CNN 和 Transformer 两个并行分支组成,用于从输入图像中提取特征,同时考虑了全局依赖性和局部信息。由于医学图像来自频谱内的某些频段,因此它们的颜色通道不像自然图像那样均匀。同时,医学分割更关注图像中的病变区域。注意力融合模块(AFM)将通道注意力和空间注意力串联起来,融合两个分支的输出特征。医学图像分割任务本质上是定位图像中物体的边界。边界增强模块(BEM)被设计在所提出网络的浅层中,以更具体地关注像素级边缘细节。三个公共数据集的实验结果验证了 CSAP-UNet 的性能优于最先进的网络,特别是在 ISIC 2017 数据集上。在Kvasir和CVC-ClinicDB上的跨数据集评估表明CSAP-UNet具有很强的泛化能力。消融实验也表明了所设计模块的有效性。训练和测试的代码可在 处获得。
更新日期:2024-03-07
down
wechat
bug