Open Access
ARTICLE
A Robust Model for Translating Arabic Sign Language into Spoken Arabic Using Deep Learning
Khalid M. O. Nahar1, Ammar Almomani2,3,*, Nahlah Shatnawi1, Mohammad Alauthman4
1 Department of Computer Sciences, Faculty of Information Technology and Computer Sciences, Yarmouk University–Irbid, 21163, Jordan
2 School of Computing, Skyline University College, Sharjah, P. O. Box 1797, United Arab Emirates
3 IT-Department-Al-Huson University College, Al-Balqa Applied University, P. O. Box 50, Irbid, Jordan
4 Department of Information Security, Faculty of Information Technology, University of Petra, Amman, Jordan
* Corresponding Authors: Ammar Almomani. Email: ,
Intelligent Automation & Soft Computing 2023, 37(2), 2037-2057. https://doi.org/10.32604/iasc.2023.038235
Received 03 December 2022; Accepted 12 April 2023; Issue published 21 June 2023
Abstract
This study presents a novel and innovative approach to automatically translating Arabic Sign Language (ATSL) into spoken Arabic. The proposed solution utilizes a deep learning-based classification approach and the transfer learning technique to retrain 12 image recognition models. The image-based translation method maps sign language gestures to corresponding letters or words using distance measures and classification as a machine learning technique. The results show that the proposed model is more accurate and faster than traditional image-based models in classifying Arabic-language signs, with a translation accuracy of 93.7%. This research makes a significant contribution to the field of ATSL. It offers a practical solution for improving communication for individuals with special needs, such as the deaf and mute community. This work demonstrates the potential of deep learning techniques in translating sign language into natural language and highlights the importance of ATSL in facilitating communication for individuals with disabilities.
Keywords
Cite This Article
APA Style
Nahar, K.M.O., Almomani, A., Shatnawi, N., Alauthman, M. (2023). A robust model for translating arabic sign language into spoken arabic using deep learning. Intelligent Automation & Soft Computing, 37(2), 2037-2057. https://doi.org/10.32604/iasc.2023.038235
Vancouver Style
Nahar KMO, Almomani A, Shatnawi N, Alauthman M. A robust model for translating arabic sign language into spoken arabic using deep learning. Intell Automat Soft Comput . 2023;37(2):2037-2057 https://doi.org/10.32604/iasc.2023.038235
IEEE Style
K.M.O. Nahar, A. Almomani, N. Shatnawi, and M. Alauthman "A Robust Model for Translating Arabic Sign Language into Spoken Arabic Using Deep Learning," Intell. Automat. Soft Comput. , vol. 37, no. 2, pp. 2037-2057. 2023. https://doi.org/10.32604/iasc.2023.038235