Open Access iconOpen Access

ARTICLE

crossmark

CGMISeg: Context-Guided Multi-Scale Interactive for Efficient Semantic Segmentation

Ze Wang, Jin Qin, Chuhua Huang*, Yongjun Zhang*

The State Key Laboratory of Public Big Data and College of Computer Science and Technology, Guizhou University, Guiyang, 550025, China

* Corresponding Authors: Chuhua Huang. Email: email; Yongjun Zhang. Email: email

(This article belongs to the Special Issue: Novel Methods for Image Classification, Object Detection, and Segmentation)

Computers, Materials & Continua 2025, 84(3), 5811-5829. https://doi.org/10.32604/cmc.2025.064537

Abstract

Semantic segmentation has made significant breakthroughs in various application fields, but achieving both accurate and efficient segmentation with limited computational resources remains a major challenge. To this end, we propose CGMISeg, an efficient semantic segmentation architecture based on a context-guided multi-scale interaction strategy, aiming to significantly reduce computational overhead while maintaining segmentation accuracy. CGMISeg consists of three core components: context-aware attention modulation, feature reconstruction, and cross-information fusion. Context-aware attention modulation is carefully designed to capture key contextual information through channel and spatial attention mechanisms. The feature reconstruction module reconstructs contextual information from different scales, modeling key rectangular areas by capturing critical contextual information in both horizontal and vertical directions, thereby enhancing the focus on foreground features. The cross-information fusion module aims to fuse the reconstructed high-level features with the original low-level features during upsampling, promoting multi-scale interaction and enhancing the model’s ability to handle objects at different scales. We extensively evaluated CGMISeg on ADE20K, Cityscapes, and COCO-Stuff, three widely used datasets benchmarks, and the experimental results show that CGMISeg exhibits significant advantages in segmentation performance, computational efficiency, and inference speed, clearly outperforming several mainstream methods, including SegFormer, Feedformer, and SegNext. Specifically, CGMISeg achieves 42.9% mIoU (Mean Intersection over Union) and 15.7 FPS (Frames Per Second) on the ADE20K dataset with 3.8 GFLOPs (Giga Floating-point Operations Per Second), outperforming Feedformer and SegNeXt by 3.7% and 1.8% in mIoU, respectively, while also offering reduced computational complexity and faster inference. CGMISeg strikes an excellent balance between accuracy and efficiency, significantly enhancing both computational and inference performance while maintaining high precision, showcasing exceptional practical value and strong potential for widespread applications.

Keywords

Semantic segmentation; context-aware attention modulation; feature reconstruction; cross-information fusion

Cite This Article

APA Style
Wang, Z., Qin, J., Huang, C., Zhang, Y. (2025). CGMISeg: Context-Guided Multi-Scale Interactive for Efficient Semantic Segmentation. Computers, Materials & Continua, 84(3), 5811–5829. https://doi.org/10.32604/cmc.2025.064537
Vancouver Style
Wang Z, Qin J, Huang C, Zhang Y. CGMISeg: Context-Guided Multi-Scale Interactive for Efficient Semantic Segmentation. Comput Mater Contin. 2025;84(3):5811–5829. https://doi.org/10.32604/cmc.2025.064537
IEEE Style
Z. Wang, J. Qin, C. Huang, and Y. Zhang, “CGMISeg: Context-Guided Multi-Scale Interactive for Efficient Semantic Segmentation,” Comput. Mater. Contin., vol. 84, no. 3, pp. 5811–5829, 2025. https://doi.org/10.32604/cmc.2025.064537



cc Copyright © 2025 The Author(s). Published by Tech Science Press.
This work is licensed under a Creative Commons Attribution 4.0 International License , which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
  • 832

    View

  • 507

    Download

  • 0

    Like

Share Link