Home / Advanced Search

  • Title/Keywords

  • Author/Affliations

  • Journal

  • Article Type

  • Start Year

  • End Year

Update SearchingClear
  • Articles
  • Online
Search Results (23)
  • Open Access

    ARTICLE

    SwinHCAD: A Robust Multi-Modality Segmentation Model for Brain Tumors Using Transformer and Channel-Wise Attention

    Seyong Jin1, Muhammad Fayaz2, L. Minh Dang3, Hyoung-Kyu Song3, Hyeonjoon Moon2,*

    CMC-Computers, Materials & Continua, Vol.86, No.1, pp. 1-23, 2026, DOI:10.32604/cmc.2025.070667 - 10 November 2025

    Abstract Brain tumors require precise segmentation for diagnosis and treatment plans due to their complex morphology and heterogeneous characteristics. While MRI-based automatic brain tumor segmentation technology reduces the burden on medical staff and provides quantitative information, existing methodologies and recent models still struggle to accurately capture and classify the fine boundaries and diverse morphologies of tumors. In order to address these challenges and maximize the performance of brain tumor segmentation, this research introduces a novel SwinUNETR-based model by integrating a new decoder block, the Hierarchical Channel-wise Attention Decoder (HCAD), into a powerful SwinUNETR encoder. The HCAD… More >

  • Open Access

    ARTICLE

    Meyer Wavelet Transform and Jaccard Deep Q Net for Small Object Classification Using Multi-Modal Images

    Mian Muhammad Kamal1,*, Syed Zain Ul Abideen2, M. A. Al-Khasawneh3,4, Alaa M. Momani4, Hala Mostafa5, Mohammed Salem Atoum6, Saeed Ullah7, Jamil Abedalrahim Jamil Alsayaydeh8,*, Mohd Faizal Bin Yusof9, Suhaila Binti Mohd Najib8

    CMES-Computer Modeling in Engineering & Sciences, Vol.144, No.3, pp. 3053-3083, 2025, DOI:10.32604/cmes.2025.067430 - 30 September 2025

    Abstract Accurate detection of small objects is critically important in high-stakes applications such as military reconnaissance and emergency rescue. However, low resolution, occlusion, and background interference make small object detection a complex and demanding task. One effective approach to overcome these issues is the integration of multimodal image data to enhance detection capabilities. This paper proposes a novel small object detection method that utilizes three types of multimodal image combinations, such as Hyperspectral–Multispectral (HS-MS), Hyperspectral–Synthetic Aperture Radar (HS-SAR), and HS-SAR–Digital Surface Model (HS-SAR-DSM). The detection process is done by the proposed Jaccard Deep Q-Net (JDQN), which More >

  • Open Access

    ARTICLE

    Face Forgery Detection via Multi-Scale Dual-Modality Mutual Enhancement Network

    Yuanqing Ding1,2, Hanming Zhai1, Qiming Ma1, Liang Zhang1, Lei Shao2, Fanliang Bu1,*

    CMC-Computers, Materials & Continua, Vol.85, No.1, pp. 905-923, 2025, DOI:10.32604/cmc.2025.066307 - 29 August 2025

    Abstract As the use of deepfake facial videos proliferate, the associated threats to social security and integrity cannot be overstated. Effective methods for detecting forged facial videos are thus urgently needed. While many deep learning-based facial forgery detection approaches show promise, they often fail to delve deeply into the complex relationships between image features and forgery indicators, limiting their effectiveness to specific forgery techniques. To address this challenge, we propose a dual-branch collaborative deepfake detection network. The network processes video frame images as input, where a specialized noise extraction module initially extracts the noise feature maps.… More >

  • Open Access

    ARTICLE

    Lip-Audio Modality Fusion for Deep Forgery Video Detection

    Yong Liu1,4, Zhiyu Wang2,*, Shouling Ji3, Daofu Gong1,5, Lanxin Cheng1, Ruosi Cheng1

    CMC-Computers, Materials & Continua, Vol.82, No.2, pp. 3499-3515, 2025, DOI:10.32604/cmc.2024.057859 - 17 February 2025

    Abstract In response to the problem of traditional methods ignoring audio modality tampering, this study aims to explore an effective deep forgery video detection technique that improves detection precision and reliability by fusing lip images and audio signals. The main method used is lip-audio matching detection technology based on the Siamese neural network, combined with MFCC (Mel Frequency Cepstrum Coefficient) feature extraction of band-pass filters, an improved dual-branch Siamese network structure, and a two-stream network structure design. Firstly, the video stream is preprocessed to extract lip images, and the audio stream is preprocessed to extract MFCC… More >

  • Open Access

    PROCEEDINGS

    Multi-Modality In-Situ Monitoring Big Data Mining for Enhanced Insight into the Laser Powder Bed Fusion Process, Structure, and Properties

    Xiayun Zhao1,*, Haolin Zhang1, Md Jahangir Alam1

    The International Conference on Computational & Experimental Engineering and Sciences, Vol.31, No.2, pp. 1-2, 2024, DOI:10.32604/icces.2024.011479

    Abstract Laser powder bed fusion (LPBF) is one predominant additive manufacturing (AM) technology for producing metallic parts with sophisticated designs that can find numerous applications in critical industries such as aerospace. To achieve precise, resilient, and intelligent LPBF, a comprehensive understanding of the dynamic processes and material responses within the actual conditions of LPBF-based AM is essential. However, obtaining such insights is challenging due to the intricate interactions among the laser, powder, part layers, and gas flow, among other factors. Multimodal in-situ monitoring is desired to visualize diverse process signatures, allowing for the direct and thorough… More >

  • Open Access

    ARTICLE

    Enhancing Multi-Modality Medical Imaging: A Novel Approach with Laplacian Filter + Discrete Fourier Transform Pre-Processing and Stationary Wavelet Transform Fusion

    Mian Muhammad Danyal1,2, Sarwar Shah Khan3,4,*, Rahim Shah Khan5, Saifullah Jan2, Naeem ur Rahman6

    Journal of Intelligent Medicine and Healthcare, Vol.2, pp. 35-53, 2024, DOI:10.32604/jimh.2024.051340 - 08 July 2024

    Abstract Multi-modality medical images are essential in healthcare as they provide valuable insights for disease diagnosis and treatment. To harness the complementary data provided by various modalities, these images are amalgamated to create a single, more informative image. This fusion process enhances the overall quality and comprehensiveness of the medical imagery, aiding healthcare professionals in making accurate diagnoses and informed treatment decisions. In this study, we propose a new hybrid pre-processing approach, Laplacian Filter + Discrete Fourier Transform (LF+DFT), to enhance medical images before fusion. The LF+DFT approach highlights key details, captures small information, and sharpens… More >

  • Open Access

    ARTICLE

    Fine-Grained Ship Recognition Based on Visible and Near-Infrared Multimodal Remote Sensing Images: Dataset, Methodology and Evaluation

    Shiwen Song, Rui Zhang, Min Hu*, Feiyao Huang

    CMC-Computers, Materials & Continua, Vol.79, No.3, pp. 5243-5271, 2024, DOI:10.32604/cmc.2024.050879 - 20 June 2024

    Abstract Fine-grained recognition of ships based on remote sensing images is crucial to safeguarding maritime rights and interests and maintaining national security. Currently, with the emergence of massive high-resolution multi-modality images, the use of multi-modality images for fine-grained recognition has become a promising technology. Fine-grained recognition of multi-modality images imposes higher requirements on the dataset samples. The key to the problem is how to extract and fuse the complementary features of multi-modality images to obtain more discriminative fusion features. The attention mechanism helps the model to pinpoint the key information in the image, resulting in a… More >

  • Open Access

    ARTICLE

    A Dual Discriminator Method for Generalized Zero-Shot Learning

    Tianshu Wei1, Jinjie Huang1,2,*

    CMC-Computers, Materials & Continua, Vol.79, No.1, pp. 1599-1612, 2024, DOI:10.32604/cmc.2024.048098 - 25 April 2024

    Abstract Zero-shot learning enables the recognition of new class samples by migrating models learned from semantic features and existing sample features to things that have never been seen before. The problems of consistency of different types of features and domain shift problems are two of the critical issues in zero-shot learning. To address both of these issues, this paper proposes a new modeling structure. The traditional approach mapped semantic features and visual features into the same feature space; based on this, a dual discriminator approach is used in the proposed model. This dual discriminator approach can… More >

  • Open Access

    ARTICLE

    Multimodality Medical Image Fusion Based on Pixel Significance with Edge-Preserving Processing for Clinical Applications

    Bhawna Goyal1, Ayush Dogra2, Dawa Chyophel Lepcha1, Rajesh Singh3, Hemant Sharma4, Ahmed Alkhayyat5, Manob Jyoti Saikia6,*

    CMC-Computers, Materials & Continua, Vol.78, No.3, pp. 4317-4342, 2024, DOI:10.32604/cmc.2024.047256 - 26 March 2024

    Abstract Multimodal medical image fusion has attained immense popularity in recent years due to its robust technology for clinical diagnosis. It fuses multiple images into a single image to improve the quality of images by retaining significant information and aiding diagnostic practitioners in diagnosing and treating many diseases. However, recent image fusion techniques have encountered several challenges, including fusion artifacts, algorithm complexity, and high computing costs. To solve these problems, this study presents a novel medical image fusion strategy by combining the benefits of pixel significance with edge-preserving processing to achieve the best fusion performance. First,… More >

  • Open Access

    ARTICLE

    Augmented Deep Multi-Granularity Pose-Aware Feature Fusion Network for Visible-Infrared Person Re-Identification

    Zheng Shi, Wanru Song*, Junhao Shan, Feng Liu

    CMC-Computers, Materials & Continua, Vol.77, No.3, pp. 3467-3488, 2023, DOI:10.32604/cmc.2023.045849 - 26 December 2023

    Abstract Visible-infrared Cross-modality Person Re-identification (VI-ReID) is a critical technology in smart public facilities such as cities, campuses and libraries. It aims to match pedestrians in visible light and infrared images for video surveillance, which poses a challenge in exploring cross-modal shared information accurately and efficiently. Therefore, multi-granularity feature learning methods have been applied in VI-ReID to extract potential multi-granularity semantic information related to pedestrian body structure attributes. However, existing research mainly uses traditional dual-stream fusion networks and overlooks the core of cross-modal learning networks, the fusion module. This paper introduces a novel network called the… More >

Displaying 1-10 on page 1 of 23. Per Page