Home / Advanced Search

  • Title/Keywords

  • Author/Affliations

  • Journal

  • Article Type

  • Start Year

  • End Year

Update SearchingClear
  • Articles
  • Online
Search Results (250)
  • Open Access

    ARTICLE

    BCCLR: A Skeleton-Based Action Recognition with Graph Convolutional Network Combining Behavior Dependence and Context Clues

    Yunhe Wang1, Yuxin Xia2, Shuai Liu2,*

    CMC-Computers, Materials & Continua, Vol.78, No.3, pp. 4489-4507, 2024, DOI:10.32604/cmc.2024.048813

    Abstract In recent years, skeleton-based action recognition has made great achievements in Computer Vision. A graph convolutional network (GCN) is effective for action recognition, modelling the human skeleton as a spatio-temporal graph. Most GCNs define the graph topology by physical relations of the human joints. However, this predefined graph ignores the spatial relationship between non-adjacent joint pairs in special actions and the behavior dependence between joint pairs, resulting in a low recognition rate for specific actions with implicit correlation between joint pairs. In addition, existing methods ignore the trend correlation between adjacent frames within an action and context clues, leading to… More >

  • Open Access

    ARTICLE

    Lightweight Cross-Modal Multispectral Pedestrian Detection Based on Spatial Reweighted Attention Mechanism

    Lujuan Deng, Ruochong Fu*, Zuhe Li, Boyi Liu, Mengze Xue, Yuhao Cui

    CMC-Computers, Materials & Continua, Vol.78, No.3, pp. 4071-4089, 2024, DOI:10.32604/cmc.2024.048200

    Abstract Multispectral pedestrian detection technology leverages infrared images to provide reliable information for visible light images, demonstrating significant advantages in low-light conditions and background occlusion scenarios. However, while continuously improving cross-modal feature extraction and fusion, ensuring the model’s detection speed is also a challenging issue. We have devised a deep learning network model for cross-modal pedestrian detection based on Resnet50, aiming to focus on more reliable features and enhance the model’s detection efficiency. This model employs a spatial attention mechanism to reweight the input visible light and infrared image data, enhancing the model’s focus on different spatial positions and sharing the… More >

  • Open Access

    ARTICLE

    Predicting Traffic Flow Using Dynamic Spatial-Temporal Graph Convolution Networks

    Yunchang Liu1,*, Fei Wan1, Chengwu Liang2

    CMC-Computers, Materials & Continua, Vol.78, No.3, pp. 4343-4361, 2024, DOI:10.32604/cmc.2024.047211

    Abstract Traffic flow prediction plays a key role in the construction of intelligent transportation system. However, due to its complex spatio-temporal dependence and its uncertainty, the research becomes very challenging. Most of the existing studies are based on graph neural networks that model traffic flow graphs and try to use fixed graph structure to deal with the relationship between nodes. However, due to the time-varying spatial correlation of the traffic network, there is no fixed node relationship, and these methods cannot effectively integrate the temporal and spatial features. This paper proposes a novel temporal-spatial dynamic graph convolutional network (TSADGCN). The dynamic… More >

  • Open Access

    ARTICLE

    Fake News Detection Based on Text-Modal Dominance and Fusing Multiple Multi-Model Clues

    Lifang Fu1, Huanxin Peng2,*, Changjin Ma2, Yuhan Liu2

    CMC-Computers, Materials & Continua, Vol.78, No.3, pp. 4399-4416, 2024, DOI:10.32604/cmc.2024.047053

    Abstract In recent years, how to efficiently and accurately identify multi-model fake news has become more challenging. First, multi-model data provides more evidence but not all are equally important. Secondly, social structure information has proven to be effective in fake news detection and how to combine it while reducing the noise information is critical. Unfortunately, existing approaches fail to handle these problems. This paper proposes a multi-model fake news detection framework based on Tex-modal Dominance and fusing Multiple Multi-model Cues (TD-MMC), which utilizes three valuable multi-model clues: text-model importance, text-image complementary, and text-image inconsistency. TD-MMC is dominated by textural content and… More >

  • Open Access

    ARTICLE

    Missing Value Imputation for Radar-Derived Time-Series Tracks of Aerial Targets Based on Improved Self-Attention-Based Network

    Zihao Song, Yan Zhou*, Wei Cheng, Futai Liang, Chenhao Zhang

    CMC-Computers, Materials & Continua, Vol.78, No.3, pp. 3349-3376, 2024, DOI:10.32604/cmc.2024.047034

    Abstract The frequent missing values in radar-derived time-series tracks of aerial targets (RTT-AT) lead to significant challenges in subsequent data-driven tasks. However, the majority of imputation research focuses on random missing (RM) that differs significantly from common missing patterns of RTT-AT. The method for solving the RM may experience performance degradation or failure when applied to RTT-AT imputation. Conventional autoregressive deep learning methods are prone to error accumulation and long-term dependency loss. In this paper, a non-autoregressive imputation model that addresses the issue of missing value imputation for two common missing patterns in RTT-AT is proposed. Our model consists of two… More >

  • Open Access

    ARTICLE

    A Cover-Independent Deep Image Hiding Method Based on Domain Attention Mechanism

    Nannan Wu1, Xianyi Chen1,*, James Msughter Adeke2, Junjie Zhao2

    CMC-Computers, Materials & Continua, Vol.78, No.3, pp. 3001-3019, 2024, DOI:10.32604/cmc.2023.045311

    Abstract Recently, deep image-hiding techniques have attracted considerable attention in covert communication and high-capacity information hiding. However, these approaches have some limitations. For example, a cover image lacks self-adaptability, information leakage, or weak concealment. To address these issues, this study proposes a universal and adaptable image-hiding method. First, a domain attention mechanism is designed by combining the Atrous convolution, which makes better use of the relationship between the secret image domain and the cover image domain. Second, to improve perceived human similarity, perceptual loss is incorporated into the training process. The experimental results are promising, with the proposed method achieving an… More >

  • Open Access

    ARTICLE

    Intelligent Fault Diagnosis Method of Rolling Bearings Based on Transfer Residual Swin Transformer with Shifted Windows

    Haomiao Wang1, Jinxi Wang2, Qingmei Sui2,*, Faye Zhang2, Yibin Li1, Mingshun Jiang2, Phanasindh Paitekul3

    Structural Durability & Health Monitoring, Vol.18, No.2, pp. 91-110, 2024, DOI:10.32604/sdhm.2023.041522

    Abstract Due to their robust learning and expression ability for complex features, the deep learning (DL) model plays a vital role in bearing fault diagnosis. However, since there are fewer labeled samples in fault diagnosis, the depth of DL models in fault diagnosis is generally shallower than that of DL models in other fields, which limits the diagnostic performance. To solve this problem, a novel transfer residual Swin Transformer (RST) is proposed for rolling bearings in this paper. RST has 24 residual self-attention layers, which use the hierarchical design and the shifted window-based residual self-attention. Combined with transfer learning techniques, the… More >

  • Open Access

    ARTICLE

    TEAM: Transformer Encoder Attention Module for Video Classification

    Hae Sung Park1, Yong Suk Choi2,*

    Computer Systems Science and Engineering, Vol.48, No.2, pp. 451-477, 2024, DOI:10.32604/csse.2023.043245

    Abstract Much like humans focus solely on object movement to understand actions, directing a deep learning model’s attention to the core contexts within videos is crucial for improving video comprehension. In the recent study, Video Masked Auto-Encoder (VideoMAE) employs a pre-training approach with a high ratio of tube masking and reconstruction, effectively mitigating spatial bias due to temporal redundancy in full video frames. This steers the model’s focus toward detailed temporal contexts. However, as the VideoMAE still relies on full video frames during the action recognition stage, it may exhibit a progressive shift in attention towards spatial contexts, deteriorating its ability… More >

  • Open Access

    ARTICLE

    CAW-YOLO: Cross-Layer Fusion and Weighted Receptive Field-Based YOLO for Small Object Detection in Remote Sensing

    Weiya Shi1,*, Shaowen Zhang2, Shiqiang Zhang2

    CMES-Computer Modeling in Engineering & Sciences, Vol.139, No.3, pp. 3209-3231, 2024, DOI:10.32604/cmes.2023.044863

    Abstract In recent years, there has been extensive research on object detection methods applied to optical remote sensing images utilizing convolutional neural networks. Despite these efforts, the detection of small objects in remote sensing remains a formidable challenge. The deep network structure will bring about the loss of object features, resulting in the loss of object features and the near elimination of some subtle features associated with small objects in deep layers. Additionally, the features of small objects are susceptible to interference from background features contained within the image, leading to a decline in detection accuracy. Moreover, the sensitivity of small… More >

  • Open Access

    ARTICLE

    Enhancing Image Description Generation through Deep Reinforcement Learning: Fusing Multiple Visual Features and Reward Mechanisms

    Yan Li, Qiyuan Wang*, Kaidi Jia

    CMC-Computers, Materials & Continua, Vol.78, No.2, pp. 2469-2489, 2024, DOI:10.32604/cmc.2024.047822

    Abstract Image description task is the intersection of computer vision and natural language processing, and it has important prospects, including helping computers understand images and obtaining information for the visually impaired. This study presents an innovative approach employing deep reinforcement learning to enhance the accuracy of natural language descriptions of images. Our method focuses on refining the reward function in deep reinforcement learning, facilitating the generation of precise descriptions by aligning visual and textual features more closely. Our approach comprises three key architectures. Firstly, it utilizes Residual Network 101 (ResNet-101) and Faster Region-based Convolutional Neural Network (Faster R-CNN) to extract average… More >

Displaying 1-10 on page 1 of 250. Per Page