
@Article{iasc.2023.035497,
AUTHOR = {Seema Sabharwal, Priti Singla},
TITLE = {Alphabet-Level Indian Sign Language Translation to Text Using Hybrid-AO Thresholding with CNN},
JOURNAL = {Intelligent Automation \& Soft Computing},
VOLUME = {37},
YEAR = {2023},
NUMBER = {3},
PAGES = {2567--2582},
URL = {http://www.techscience.com/iasc/v37n3/54116},
ISSN = {2326-005X},
ABSTRACT = {Sign language is used as a communication medium in the field of trade, defence, and in deaf-mute communities worldwide. Over the last few decades, research in the domain of translation of sign language has grown and become more challenging. This necessitates the development of a Sign Language Translation System (SLTS) to provide effective communication in different research domains. In this paper, novel Hybrid Adaptive Gaussian Thresholding with Otsu Algorithm (Hybrid-AO) for image segmentation is proposed for the translation of alphabet-level Indian Sign Language (ISLTS) with a 5-layer Convolution Neural Network (CNN). The focus of this paper is to analyze various image segmentation (Canny Edge Detection, Simple Thresholding, and Hybrid-AO), pooling approaches (Max, Average, and Global Average Pooling), and activation functions (ReLU, Leaky ReLU, and ELU). 5-layer CNN with Max pooling, Leaky ReLU activation function, and Hybrid-AO (5MXLR-HAO) have outperformed other frameworks. An open-access dataset of ISL alphabets with approx. 31 K images of 26 classes have been used to train and test the model. The proposed framework has been developed for translating alphabet-level Indian Sign Language into text. The proposed framework attains 98.95% training accuracy, 98.05% validation accuracy, and 0.0721 training loss and 0.1021 validation loss and the performance of the proposed system outperforms other existing systems.},
DOI = {10.32604/iasc.2023.035497}
}



