Open Access iconOpen Access

ARTICLE

CMACF-Net: Cross-Multiscale Adaptive Collaborative and Fusion Grasp Detection Network

Xi Li1,2, Runpu Nie1,*, Zhaoyong Fan2, Lianying Zou2, Zhenhua Xiao2, Kaile Dong1

1 School of Electrical and Information Engineering, Wuhan Institute of Technology, Wuhan, 430205, China
2 College of Information and Artificial Intelligence, Nanchang Institute of Science and Technology, Nanchang, 330108, China

* Corresponding Author: Runpu Nie. Email: email

Computers, Materials & Continua 2025, 85(2), 2959-2984. https://doi.org/10.32604/cmc.2025.066740

Abstract

With the rapid development of robotics, grasp prediction has become fundamental to achieving intelligent physical interactions. To enhance grasp detection accuracy in unstructured environments, we propose a novel Cross-Multiscale Adaptive Collaborative and Fusion Grasp Detection Network (CMACF-Net). Addressing the limitations of conventional methods in capturing multi-scale spatial features, CMACF-Net introduces the Quantized Multi-scale Global Attention Module (QMGAM), which enables precise multi-scale spatial calibration and adaptive spatial-channel interaction, ultimately yielding a more robust and discriminative feature representation. To reduce the degradation of local features and the loss of high-frequency information, the Cross-scale Context Integration Module (CCI) is employed to facilitate the effective fusion and alignment of global context and local details. Furthermore, an Efficient Up-Convolution Block (EUCB) is integrated into a U-Net architecture to effectively restore spatial details lost during the downsampling process, while simultaneously preserving computational efficiency. Extensive evaluations demonstrate that CMACF-Net achieves state-of-the-art detection accuracies of 98.9% and 95.9% on the Cornell and Jacquard datasets, respectively. Additionally, real-time grasping experiments on the RM65-B robotic platform validate the framework’s robustness and generalization capability, underscoring its applicability to real-world robotic manipulation scenarios.

Keywords

Robot grasp; grasp detection; convolutional neural network; vision transformer; attention mechanism

Cite This Article

APA Style
Li, X., Nie, R., Fan, Z., Zou, L., Xiao, Z. et al. (2025). CMACF-Net: Cross-Multiscale Adaptive Collaborative and Fusion Grasp Detection Network. Computers, Materials & Continua, 85(2), 2959–2984. https://doi.org/10.32604/cmc.2025.066740
Vancouver Style
Li X, Nie R, Fan Z, Zou L, Xiao Z, Dong K. CMACF-Net: Cross-Multiscale Adaptive Collaborative and Fusion Grasp Detection Network. Comput Mater Contin. 2025;85(2):2959–2984. https://doi.org/10.32604/cmc.2025.066740
IEEE Style
X. Li, R. Nie, Z. Fan, L. Zou, Z. Xiao, and K. Dong, “CMACF-Net: Cross-Multiscale Adaptive Collaborative and Fusion Grasp Detection Network,” Comput. Mater. Contin., vol. 85, no. 2, pp. 2959–2984, 2025. https://doi.org/10.32604/cmc.2025.066740



cc Copyright © 2025 The Author(s). Published by Tech Science Press.
This work is licensed under a Creative Commons Attribution 4.0 International License , which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
  • 1110

    View

  • 626

    Download

  • 0

    Like

Share Link