Jianuo Huang1,2, Bohan Lai2, Weiye Qiu3, Caixu Xu4, Jie He1,5,*
CMC-Computers, Materials & Continua, Vol.82, No.3, pp. 4841-4862, 2025, DOI:10.32604/cmc.2025.059733
- 06 March 2025
Abstract The self-attention mechanism of Transformers, which captures long-range contextual information, has demonstrated significant potential in image segmentation. However, their ability to learn local, contextual relationships between pixels requires further improvement. Previous methods face challenges in efficiently managing multi-scale features of different granularities from the encoder backbone, leaving room for improvement in their global representation and feature extraction capabilities. To address these challenges, we propose a novel Decoder with Multi-Head Feature Receptors (DMHFR), which receives multi-scale features from the encoder backbone and organizes them into three feature groups with different granularities: coarse, fine-grained, and full set.… More >