Open Access
ARTICLE
Enhanced BEV Scene Segmentation: De-Noise Channel Attention for Resource-Constrained Environments
1 College of Computer Science, Chongqing University, Chongqing, 400044, China
2 SUGON Industrial Control and Security Center, Chengdu, 610225, China
3 Department of Computer Science, Maharishi University of Management, Fairfield, IA 52557, USA
* Corresponding Author: Yunfei Yin. Email:
Computers, Materials & Continua 2026, 87(1), 90 https://doi.org/10.32604/cmc.2025.074122
Received 02 October 2025; Accepted 25 December 2025; Issue published 10 February 2026
Abstract
Autonomous vehicles rely heavily on accurate and efficient scene segmentation for safe navigation and efficient operations. Traditional Bird’s Eye View (BEV) methods on semantic scene segmentation, which leverage multimodal sensor fusion, often struggle with noisy data and demand high-performance GPUs, leading to sensor misalignment and performance degradation. This paper introduces an Enhanced Channel Attention BEV (ECABEV), a novel approach designed to address the challenges under insufficient GPU memory conditions. ECABEV integrates camera and radar data through a de-noise enhanced channel attention mechanism, which utilizes global average and max pooling to effectively filter out noise while preserving discriminative features. Furthermore, an improved fusion approach is proposed to efficiently merge categorical data across modalities. To reduce computational overhead, a bilinear interpolation layer normalization method is devised to ensure spatial feature fidelity. Moreover, a scalable cross-entropy loss function is further designed to handle the imbalanced classes with less computational efficiency sacrifice. Extensive experiments on the nuScenes dataset demonstrate that ECABEV achieves state-of-the-art performance with an IoU of 39.961, using a lightweight ViT-B/14 backbone and lower resolution (224 × 224). Our approach highlights its cost-effectiveness and practical applicability, even on low-end devices. The code is publicly available at: https://github.com/YYF-CQU/ECABEV.git.Keywords
Cite This Article
Copyright © 2026 The Author(s). Published by Tech Science Press.This work is licensed under a Creative Commons Attribution 4.0 International License , which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.


Submit a Paper
Propose a Special lssue
View Full Text
Download PDF
Downloads
Citation Tools