Home / Advanced Search

  • Title/Keywords

  • Author/Affliations

  • Journal

  • Article Type

  • Start Year

  • End Year

Update SearchingClear
  • Articles
  • Online
Search Results (48)
  • Open Access

    ARTICLE

    Enhancing Classroom Behavior Recognition with Lightweight Multi-Scale Feature Fusion

    Chuanchuan Wang1,2, Ahmad Sufril Azlan Mohamed2,*, Xiao Yang 2, Hao Zhang 2, Xiang Li1, Mohd Halim Bin Mohd Noor 2

    CMC-Computers, Materials & Continua, Vol.85, No.1, pp. 855-874, 2025, DOI:10.32604/cmc.2025.066343 - 29 August 2025

    Abstract Classroom behavior recognition is a hot research topic, which plays a vital role in assessing and improving the quality of classroom teaching. However, existing classroom behavior recognition methods have challenges for high recognition accuracy with datasets with problems such as scenes with blurred pictures, and inconsistent objects. To address this challenge, we proposed an effective, lightweight object detector method called the RFNet model (YOLO-FR). The YOLO-FR is a lightweight and effective model. Specifically, for efficient multi-scale feature extraction, effective feature pyramid shared convolutional (FPSC) was designed to improve the feature extract performance by leveraging convolutional… More >

  • Open Access

    ARTICLE

    A YOLOv11-Based Deep Learning Framework for Multi-Class Human Action Recognition

    Nayeemul Islam Nayeem1, Shirin Mahbuba1, Sanjida Islam Disha1, Md Rifat Hossain Buiyan1, Shakila Rahman1,*, M. Abdullah-Al-Wadud2, Jia Uddin3,*

    CMC-Computers, Materials & Continua, Vol.85, No.1, pp. 1541-1557, 2025, DOI:10.32604/cmc.2025.065061 - 29 August 2025

    Abstract Human activity recognition is a significant area of research in artificial intelligence for surveillance, healthcare, sports, and human-computer interaction applications. The article benchmarks the performance of You Only Look Once version 11-based (YOLOv11-based) architecture for multi-class human activity recognition. The article benchmarks the performance of You Only Look Once version 11-based (YOLOv11-based) architecture for multi-class human activity recognition. The dataset consists of 14,186 images across 19 activity classes, from dynamic activities such as running and swimming to static activities such as sitting and sleeping. Preprocessing included resizing all images to 512 512 pixels, annotating them… More >

  • Open Access

    ARTICLE

    A Novel Attention-Based Parallel Blocks Deep Architecture for Human Action Recognition

    Yasir Khan Jadoon1, Yasir Noman Khalid1, Muhammad Attique Khan2, Jungpil Shin3,*, Fatimah Alhayan4, Hee-Chan Cho5, Byoungchol Chang6,*

    CMES-Computer Modeling in Engineering & Sciences, Vol.144, No.1, pp. 1143-1164, 2025, DOI:10.32604/cmes.2025.066984 - 31 July 2025

    Abstract Real-time surveillance is attributed to recognizing the variety of actions performed by humans. Human Action Recognition (HAR) is a technique that recognizes human actions from a video stream. A range of variations in human actions makes it difficult to recognize with considerable accuracy. This paper presents a novel deep neural network architecture called Attention RB-Net for HAR using video frames. The input is provided to the model in the form of video frames. The proposed deep architecture is based on the unique structuring of residual blocks with several filter sizes. Features are extracted from each… More >

  • Open Access

    ARTICLE

    ARNet: Integrating Spatial and Temporal Deep Learning for Robust Action Recognition in Videos

    Hussain Dawood1, Marriam Nawaz2, Tahira Nazir3, Ali Javed2, Abdul Khader Jilani Saudagar4,*, Hatoon S. AlSagri4

    CMES-Computer Modeling in Engineering & Sciences, Vol.144, No.1, pp. 429-459, 2025, DOI:10.32604/cmes.2025.066415 - 31 July 2025

    Abstract Reliable human action recognition (HAR) in video sequences is critical for a wide range of applications, such as security surveillance, healthcare monitoring, and human-computer interaction. Several automated systems have been designed for this purpose; however, existing methods often struggle to effectively integrate spatial and temporal information from input samples such as 2-stream networks or 3D convolutional neural networks (CNNs), which limits their accuracy in discriminating numerous human actions. Therefore, this study introduces a novel deep-learning framework called the ARNet, designed for robust HAR. ARNet consists of two main modules, namely, a refined InceptionResNet-V2-based CNN and… More >

  • Open Access

    ARTICLE

    Prediction of Assembly Intent for Human-Robot Collaboration Based on Video Analytics and Hidden Markov Model

    Jing Qu1, Yanmei Li1,2, Changrong Liu1, Wen Wang1, Weiping Fu1,3,*

    CMC-Computers, Materials & Continua, Vol.84, No.2, pp. 3787-3810, 2025, DOI:10.32604/cmc.2025.065895 - 03 July 2025

    Abstract Despite the gradual transformation of traditional manufacturing by the Human-Robot Collaboration Assembly (HRCA), challenges remain in the robot’s ability to understand and predict human assembly intentions. This study aims to enhance the robot’s comprehension and prediction capabilities of operator assembly intentions by capturing and analyzing operator behavior and movements. We propose a video feature extraction method based on the Temporal Shift Module Network (TSM-ResNet50) to extract spatiotemporal features from assembly videos and differentiate various assembly actions using feature differences between video frames. Furthermore, we construct an action recognition and segmentation model based on the Refined-Multi-Scale… More >

  • Open Access

    ARTICLE

    Video Action Recognition Method Based on Personalized Federated Learning and Spatiotemporal Features

    Rongsen Wu1, Jie Xu1, Yuhang Zhang1, Changming Zhao2,*, Yiweng Xie3, Zelei Wu1, Yunji Li2, Jinhong Guo4, Shiyang Tang5,6

    CMC-Computers, Materials & Continua, Vol.83, No.3, pp. 4961-4978, 2025, DOI:10.32604/cmc.2025.061396 - 19 May 2025

    Abstract With the rapid development of artificial intelligence and Internet of Things technologies, video action recognition technology is widely applied in various scenarios, such as personal life and industrial production. However, while enjoying the convenience brought by this technology, it is crucial to effectively protect the privacy of users’ video data. Therefore, this paper proposes a video action recognition method based on personalized federated learning and spatiotemporal features. Under the framework of federated learning, a video action recognition method leveraging spatiotemporal features is designed. For the local spatiotemporal features of the video, a new differential information… More >

  • Open Access

    ARTICLE

    Lightweight Classroom Student Action Recognition Method Based on Spatiotemporal Multimodal Feature Fusion

    Shaodong Zou1, Di Wu1, Jianhou Gan1,2,*, Juxiang Zhou1,2, Jiatian Mei1,2

    CMC-Computers, Materials & Continua, Vol.83, No.1, pp. 1101-1116, 2025, DOI:10.32604/cmc.2025.061376 - 26 March 2025

    Abstract The task of student action recognition in the classroom is to precisely capture and analyze the actions of students in classroom videos, providing a foundation for realizing intelligent and accurate teaching. However, the complex nature of the classroom environment has added challenges and difficulties in the process of student action recognition. In this research article, with regard to the circumstances where students are prone to be occluded and classroom computing resources are restricted in real classroom scenarios, a lightweight multi-modal fusion action recognition approach is put forward. This proposed method is capable of enhancing the… More >

  • Open Access

    ARTICLE

    Skeleton-Based Action Recognition Using Graph Convolutional Network with Pose Correction and Channel Topology Refinement

    Yuxin Gao1, Xiaodong Duan2,3, Qiguo Dai2,3,*

    CMC-Computers, Materials & Continua, Vol.83, No.1, pp. 701-718, 2025, DOI:10.32604/cmc.2025.060137 - 26 March 2025

    Abstract Graph convolutional network (GCN) as an essential tool in human action recognition tasks have achieved excellent performance in previous studies. However, most current skeleton-based action recognition using GCN methods use a shared topology, which cannot flexibly adapt to the diverse correlations between joints under different motion features. The video-shooting angle or the occlusion of the body parts may bring about errors when extracting the human pose coordinates with estimation algorithms. In this work, we propose a novel graph convolutional learning framework, called PCCTR-GCN, which integrates pose correction and channel topology refinement for skeleton-based human action… More >

  • Open Access

    ARTICLE

    Improving Badminton Action Recognition Using Spatio-Temporal Analysis and a Weighted Ensemble Learning Model

    Farida Asriani1,2, Azhari Azhari1,*, Wahyono Wahyono1

    CMC-Computers, Materials & Continua, Vol.81, No.2, pp. 3079-3096, 2024, DOI:10.32604/cmc.2024.058193 - 18 November 2024

    Abstract Incredible progress has been made in human action recognition (HAR), significantly impacting computer vision applications in sports analytics. However, identifying dynamic and complex movements in sports like badminton remains challenging due to the need for precise recognition accuracy and better management of complex motion patterns. Deep learning techniques like convolutional neural networks (CNNs), long short-term memory (LSTM), and graph convolutional networks (GCNs) improve recognition in large datasets, while the traditional machine learning methods like SVM (support vector machines), RF (random forest), and LR (logistic regression), combined with handcrafted features and ensemble approaches, perform well but… More >

  • Open Access

    ARTICLE

    Robust Human Interaction Recognition Using Extended Kalman Filter

    Tanvir Fatima Naik Bukht1, Abdulwahab Alazeb2, Naif Al Mudawi2, Bayan Alabdullah3, Khaled Alnowaiser4, Ahmad Jalal1, Hui Liu5,*

    CMC-Computers, Materials & Continua, Vol.81, No.2, pp. 2987-3002, 2024, DOI:10.32604/cmc.2024.053547 - 18 November 2024

    Abstract In the field of computer vision and pattern recognition, knowledge based on images of human activity has gained popularity as a research topic. Activity recognition is the process of determining human behavior based on an image. We implemented an Extended Kalman filter to create an activity recognition system here. The proposed method applies an HSI color transformation in its initial stages to improve the clarity of the frame of the image. To minimize noise, we use Gaussian filters. Extraction of silhouette using the statistical method. We use Binary Robust Invariant Scalable Keypoints (BRISK) and SIFT More >

Displaying 1-10 on page 1 of 48. Per Page