TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Multi-scale self-guided attention for medical image segmen...

Multi-scale self-guided attention for medical image segmentation

Ashish Sinha, Jose Dolz

2019-06-07arXiv preprint 2019 6Deep AttentionSegmentationSemantic SegmentationMedical Image SegmentationBrain Tumor SegmentationImage Segmentation
PaperPDFCode(official)

Abstract

Even though convolutional neural networks (CNNs) are driving progress in medical image segmentation, standard models still have some drawbacks. First, the use of multi-scale approaches, i.e., encoder-decoder architectures, leads to a redundant use of information, where similar low-level features are extracted multiple times at multiple scales. Second, long-range feature dependencies are not efficiently modeled, resulting in non-optimal discriminative feature representations associated with each semantic class. In this paper we attempt to overcome these limitations with the proposed architecture, by capturing richer contextual dependencies based on the use of guided self-attention mechanisms. This approach is able to integrate local features with their corresponding global dependencies, as well as highlight interdependent channel maps in an adaptive manner. Further, the additional loss between different modules guides the attention mechanisms to neglect irrelevant information and focus on more discriminant regions of the image by emphasizing relevant feature associations. We evaluate the proposed model in the context of semantic segmentation on three different datasets: abdominal organs, cardiovascular structures and brain tumors. A series of ablation experiments support the importance of these attention modules in the proposed architecture. In addition, compared to other state-of-the-art segmentation networks our model yields better segmentation performance, increasing the accuracy of the predictions while reducing the standard deviation. This demonstrates the efficiency of our approach to generate precise and reliable automatic segmentations of medical images. Our code is made publicly available at https://github.com/sinAshish/Multi-Scale-Attention

Results

TaskDatasetMetricValueModel
Medical Image SegmentationHSVMDice Score83.2MS-Dual-Guided
Medical Image SegmentationHSVMMSD1.19MS-Dual-Guided
Medical Image SegmentationHSVMVS94.45MS-Dual-Guided
Medical Image SegmentationCHAOS MRI DatasetDice Score86.75MS-Dual-Guided
Medical Image SegmentationCHAOS MRI DatasetMSD66MS-Dual-Guided
Medical Image SegmentationCHAOS MRI DatasetVS93.85MS-Dual-Guided
Medical Image SegmentationBRATS 2018Dice Score0.8037MS-Dual-Guided
Medical Image SegmentationBRATS 2018MSD0.9MS-Dual-Guided
Medical Image SegmentationBRATS 2018VS93.08MS-Dual-Guided

Related Papers

SeC: Advancing Complex Video Object Segmentation via Progressive Concept Construction2025-07-21Deep Learning-Based Fetal Lung Segmentation from Diffusion-weighted MRI Images and Lung Maturity Evaluation for Fetal Growth Restriction2025-07-17DiffOSeg: Omni Medical Image Segmentation via Multi-Expert Collaboration Diffusion Model2025-07-17From Variability To Accuracy: Conditional Bernoulli Diffusion Models with Consensus-Driven Correction for Thin Structure Segmentation2025-07-17Unleashing Vision Foundation Models for Coronary Artery Segmentation: Parallel ViT-CNN Encoding and Variational Fusion2025-07-17SCORE: Scene Context Matters in Open-Vocabulary Remote Sensing Instance Segmentation2025-07-17Unified Medical Image Segmentation with State Space Modeling Snake2025-07-17A Privacy-Preserving Semantic-Segmentation Method Using Domain-Adaptation Technique2025-07-17