File Download

There are no files associated with this item.

  Links for fulltext
     (May Require Subscription)
Supplementary

Article: A novel full-convolution UNet-transformer for medical image segmentation

TitleA novel full-convolution UNet-transformer for medical image segmentation
Authors
KeywordsDepth-wise convolutions
Global–local attention modules
Image segmentation
Medical images
Reparametrized feedforward networks
Issue Date2024
Citation
Biomedical Signal Processing and Control, 2024, v. 89, article no. 105772 How to Cite?
AbstractThe Transformer-based methods are still unable to effectively model local contexts although they make up for the deficiency of remote information dependencies for approaches based on small kernel CNNs. To overcome such a shortage, this paper proposes a novel full-convolution UNet Transformer model, FC-UNETTR, for medical image segmentation. First, a novel global–local attention module is proposed by utilizing multiple small kernels of different sizes for depth-wise convolutions to expand the receptive field of the network model, increase the remote dependence of semantic information in the encoder stage, and also improve the feature extraction capability of the network for fuzzy edges. Then, a reparametrized feedforward network is developed to further improve the local information extraction and mitigate the coupling between feature maps such that the relationship between feature map channels can be better revealed. Furthermore, the skip connection and decoder are redesigned by constructing a dense multiscale module instead of traditional ResNet modules to mitigate semantic bias. Benefiting from the above improvements, the constructed FC-UNETTR without pre-training demonstrates a strong capability to extract local features and capture long-range dependencies of images in medical image segmentation. Experiments show that FC-UNETTR achieves an excellent performance of 85.67% for DSC and 7.82 for HD metrics on the Synapse dataset with fewer model parameters compared with state-of-the-art networks. Furthermore, DSC reaches 92.46% and 94.76% on the ACDC dataset and the private dataset of oral graft bone, respectively, outperforming some of the latest medical image segmentation models.
Persistent Identifierhttp://hdl.handle.net/10722/354398
ISSN
2023 Impact Factor: 4.9
2023 SCImago Journal Rankings: 1.284
ISI Accession Number ID

 

DC FieldValueLanguage
dc.contributor.authorZhu, Tianyou-
dc.contributor.authorDing, Derui-
dc.contributor.authorWang, Feng-
dc.contributor.authorLiang, Wei-
dc.contributor.authorWang, Bo-
dc.date.accessioned2025-02-07T08:48:21Z-
dc.date.available2025-02-07T08:48:21Z-
dc.date.issued2024-
dc.identifier.citationBiomedical Signal Processing and Control, 2024, v. 89, article no. 105772-
dc.identifier.issn1746-8094-
dc.identifier.urihttp://hdl.handle.net/10722/354398-
dc.description.abstractThe Transformer-based methods are still unable to effectively model local contexts although they make up for the deficiency of remote information dependencies for approaches based on small kernel CNNs. To overcome such a shortage, this paper proposes a novel full-convolution UNet Transformer model, FC-UNETTR, for medical image segmentation. First, a novel global–local attention module is proposed by utilizing multiple small kernels of different sizes for depth-wise convolutions to expand the receptive field of the network model, increase the remote dependence of semantic information in the encoder stage, and also improve the feature extraction capability of the network for fuzzy edges. Then, a reparametrized feedforward network is developed to further improve the local information extraction and mitigate the coupling between feature maps such that the relationship between feature map channels can be better revealed. Furthermore, the skip connection and decoder are redesigned by constructing a dense multiscale module instead of traditional ResNet modules to mitigate semantic bias. Benefiting from the above improvements, the constructed FC-UNETTR without pre-training demonstrates a strong capability to extract local features and capture long-range dependencies of images in medical image segmentation. Experiments show that FC-UNETTR achieves an excellent performance of 85.67% for DSC and 7.82 for HD metrics on the Synapse dataset with fewer model parameters compared with state-of-the-art networks. Furthermore, DSC reaches 92.46% and 94.76% on the ACDC dataset and the private dataset of oral graft bone, respectively, outperforming some of the latest medical image segmentation models.-
dc.languageeng-
dc.relation.ispartofBiomedical Signal Processing and Control-
dc.subjectDepth-wise convolutions-
dc.subjectGlobal–local attention modules-
dc.subjectImage segmentation-
dc.subjectMedical images-
dc.subjectReparametrized feedforward networks-
dc.titleA novel full-convolution UNet-transformer for medical image segmentation-
dc.typeArticle-
dc.description.naturelink_to_subscribed_fulltext-
dc.identifier.doi10.1016/j.bspc.2023.105772-
dc.identifier.scopuseid_2-s2.0-85179136839-
dc.identifier.volume89-
dc.identifier.spagearticle no. 105772-
dc.identifier.epagearticle no. 105772-
dc.identifier.eissn1746-8108-
dc.identifier.isiWOS:001135726500001-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats