Home / Advanced Search

  • Title/Keywords

  • Author/Affliations

  • Journal

  • Article Type

  • Start Year

  • End Year

Update SearchingClear
  • Articles
  • Online
Search Results (29)
  • Open Access

    ARTICLE

    Automated Speech Recognition System to Detect Babies’ Feelings through Feature Analysis

    Sana Yasin1, Umar Draz2,3,*, Tariq Ali4, Kashaf Shahid1, Amna Abid1, Rukhsana Bibi1, Muhammad Irfan5, Mohammed A. Huneif6, Sultan A. Almedhesh6, Seham M. Alqahtani6, Alqahtani Abdulwahab6, Mohammed Jamaan Alzahrani6, Dhafer Batti Alshehri6, Alshehri Ali Abdullah7, Saifur Rahman5

    CMC-Computers, Materials & Continua, Vol.73, No.2, pp. 4349-4367, 2022, DOI:10.32604/cmc.2022.028251

    Abstract Diagnosing a baby’s feelings poses a challenge for both doctors and parents because babies cannot explain their feelings through expression or speech. Understanding the emotions of babies and their associated expressions during different sensations such as hunger, pain, etc., is a complicated task. In infancy, all communication and feelings are propagated through cry-speech, which is a natural phenomenon. Several clinical methods can be used to diagnose a baby’s diseases, but nonclinical methods of diagnosing a baby’s feelings are lacking. As such, in this study, we aimed to identify babies’ feelings and emotions through their cry using a nonclinical method. Changes… More >

  • Open Access

    ARTICLE

    Cross-Language Transfer Learning-based Lhasa-Tibetan Speech Recognition

    Zhijie Wang1, Yue Zhao1,*, Licheng Wu1, Xiaojun Bi1, Zhuoma Dawa2, Qiang Ji3

    CMC-Computers, Materials & Continua, Vol.73, No.1, pp. 629-639, 2022, DOI:10.32604/cmc.2022.027092

    Abstract As one of Chinese minority languages, Tibetan speech recognition technology was not researched upon as extensively as Chinese and English were until recently. This, along with the relatively small Tibetan corpus, has resulted in an unsatisfying performance of Tibetan speech recognition based on an end-to-end model. This paper aims to achieve an accurate Tibetan speech recognition using a small amount of Tibetan training data. We demonstrate effective methods of Tibetan end-to-end speech recognition via cross-language transfer learning from three aspects: modeling unit selection, transfer learning method, and source language selection. Experimental results show that the Chinese-Tibetan multi-language learning method using… More >

  • Open Access

    ARTICLE

    Speak-Correct: A Computerized Interface for the Analysis of Mispronounced Errors

    Kamal Jambi1,*, Hassanin Al-Barhamtoshy1, Wajdi Al-Jedaibi1, Mohsen Rashwan2, Sherif Abdou3

    Computer Systems Science and Engineering, Vol.43, No.3, pp. 1155-1173, 2022, DOI:10.32604/csse.2022.024967

    Abstract Any natural language may have dozens of accents. Even though the equivalent phonemic formation of the word, if it is properly called in different accents, humans do have audio signals that are distinct from one another. Among the most common issues with speech, the processing is discrepancies in pronunciation, accent, and enunciation. This research study examines the issues of detecting, fixing, and summarising accent defects of average Arabic individuals in English-speaking speech. The article then discusses the key approaches and structure that will be utilized to address both accent flaws and pronunciation issues. The proposed SpeakCorrect computerized interface employs a… More >

  • Open Access

    ARTICLE

    An Innovative Approach Utilizing Binary-View Transformer for Speech Recognition Task

    Muhammad Babar Kamal1, Arfat Ahmad Khan2, Faizan Ahmed Khan3, Malik Muhammad Ali Shahid4, Chitapong Wechtaisong2,*, Muhammad Daud Kamal5, Muhammad Junaid Ali6, Peerapong Uthansakul2

    CMC-Computers, Materials & Continua, Vol.72, No.3, pp. 5547-5562, 2022, DOI:10.32604/cmc.2022.024590

    Abstract The deep learning advancements have greatly improved the performance of speech recognition systems, and most recent systems are based on the Recurrent Neural Network (RNN). Overall, the RNN works fine with the small sequence data, but suffers from the gradient vanishing problem in case of large sequence. The transformer networks have neutralized this issue and have shown state-of-the-art results on sequential or speech-related data. Generally, in speech recognition, the input audio is converted into an image using Mel-spectrogram to illustrate frequencies and intensities. The image is classified by the machine learning mechanism to generate a classification transcript. However, the audio… More >

  • Open Access

    ARTICLE

    Enhanced Marathi Speech Recognition Facilitated by Grasshopper Optimisation-Based Recurrent Neural Network

    Ravindra Parshuram Bachate1, Ashok Sharma2, Amar Singh3, Ayman A. Aly4, Abdulaziz H. Alghtani4, Dac-Nhuong Le5,6,*

    Computer Systems Science and Engineering, Vol.43, No.2, pp. 439-454, 2022, DOI:10.32604/csse.2022.024214

    Abstract Communication is a significant part of being human and living in the world. Diverse kinds of languages and their variations are there; thus, one person can speak any language and cannot effectively communicate with one who speaks that language in a different accent. Numerous application fields such as education, mobility, smart systems, security, and health care systems utilize the speech or voice recognition models abundantly. Though, various studies are focused on the Arabic or Asian and English languages by ignoring other significant languages like Marathi that leads to the broader research motivations in regional languages. It is necessary to understand… More >

  • Open Access

    ARTICLE

    Hybrid In-Vehicle Background Noise Reduction for Robust Speech Recognition: The Possibilities of Next Generation 5G Data Networks

    Radek Martinek1, Jan Baros1, Rene Jaros1, Lukas Danys1,*, Jan Nedoma2

    CMC-Computers, Materials & Continua, Vol.71, No.3, pp. 4659-4676, 2022, DOI:10.32604/cmc.2022.019904

    Abstract This pilot study focuses on employment of hybrid LMS-ICA system for in-vehicle background noise reduction. Modern vehicles are nowadays increasingly supporting voice commands, which are one of the pillars of autonomous and SMART vehicles. Robust speaker recognition for context-aware in-vehicle applications is limited to a certain extent by in-vehicle background noise. This article presents the new concept of a hybrid system, which is implemented as a virtual instrument. The highly modular concept of the virtual car used in combination with real recordings of various driving scenarios enables effective testing of the investigated methods of in-vehicle background noise reduction. The study… More >

  • Open Access

    ARTICLE

    End-to-End Speech Recognition of Tamil Language

    Mohamed Hashim Changrampadi1,*, A. Shahina2, M. Badri Narayanan2, A. Nayeemulla Khan3

    Intelligent Automation & Soft Computing, Vol.32, No.2, pp. 1309-1323, 2022, DOI:10.32604/iasc.2022.022021

    Abstract Research in speech recognition is progressing with numerous state-of-the-art results in recent times. However, relatively fewer research is being carried out in Automatic Speech Recognition (ASR) for languages with low resources. We present a method to develop speech recognition model with minimal resources using Mozilla DeepSpeech architecture. We have utilized freely available online computational resources for training, enabling similar approaches to be carried out for research in a low-resourced languages in a financially constrained environments. We also present novel ways to build an efficient language model from publicly available web resources to improve accuracy in ASR. The proposed ASR model… More >

  • Open Access

    ARTICLE

    Deep Learning-Based Approach for Arabic Visual Speech Recognition

    Nadia H. Alsulami1,*, Amani T. Jamal1, Lamiaa A. Elrefaei2

    CMC-Computers, Materials & Continua, Vol.71, No.1, pp. 85-108, 2022, DOI:10.32604/cmc.2022.019450

    Abstract Lip-reading technologies are rapidly progressing following the breakthrough of deep learning. It plays a vital role in its many applications, such as: human-machine communication practices or security applications. In this paper, we propose to develop an effective lip-reading recognition model for Arabic visual speech recognition by implementing deep learning algorithms. The Arabic visual datasets that have been collected contains 2400 records of Arabic digits and 960 records of Arabic phrases from 24 native speakers. The primary purpose is to provide a high-performance model in terms of enhancing the preprocessing phase. Firstly, we extract keyframes from our dataset. Secondly, we produce… More >

  • Open Access

    ARTICLE

    Speech Recognition-Based Automated Visual Acuity Testing with Adaptive Mel Filter Bank

    Shibli Nisar1, Muhammad Asghar Khan2,*, Fahad Algarni3, Abdul Wakeel1, M. Irfan Uddin4, Insaf Ullah2

    CMC-Computers, Materials & Continua, Vol.70, No.2, pp. 2991-3004, 2022, DOI:10.32604/cmc.2022.020376

    Abstract One of the most commonly reported disabilities is vision loss, which can be diagnosed by an ophthalmologist in order to determine the visual system of a patient. This procedure, however, usually requires an appointment with an ophthalmologist, which is both time-consuming and expensive process. Other issues that can arise include a lack of appropriate equipment and trained practitioners, especially in rural areas. Centered on a cognitively motivated attribute extraction and speech recognition approach, this paper proposes a novel idea that immediately determines the eyesight deficiency. The proposed system uses an adaptive filter bank with weighted mel frequency cepstral coefficients for… More >

  • Open Access

    ARTICLE

    Noise Reduction in Industry Based on Virtual Instrumentation

    Radek Martinek1, Rene Jaros1, Jan Baros1, Lukas Danys1, Aleksandra Kawala-Sterniuk2, Jan Nedoma3,*, Zdenek Machacek1, Jiri Koziorek1

    CMC-Computers, Materials & Continua, Vol.69, No.1, pp. 1073-1096, 2021, DOI:10.32604/cmc.2021.017568

    Abstract This paper discusses the reduction of background noise in an industrial environment to extend human-machine-interaction. In the Industry 4.0 era, the mass development of voice control (speech recognition) in various industrial applications is possible, especially as related to augmented reality (such as hands-free control via voice commands). As Industry 4.0 relies heavily on radiofrequency technologies, some brief insight into this problem is provided, including the Internet of things (IoT) and 5G deployment. This study was carried out in cooperation with the industrial partner Brose CZ spol. s.r.o., where sound recordings were made to produce a dataset. The experimental environment comprised… More >

Displaying 11-20 on page 2 of 29. Per Page