@Article{cmc.2023.034072, AUTHOR = {Y. C. A. Padmanabha Reddy, Shyam Sunder Reddy Kasireddy, Nageswara Rao Sirisala, Ramu Kuchipudi, Purnachand Kollapudi}, TITLE = {An Efficient Long Short-Term Memory Model for Digital Cross-Language Summarization}, JOURNAL = {Computers, Materials \& Continua}, VOLUME = {74}, YEAR = {2023}, NUMBER = {3}, PAGES = {6389--6409}, URL = {http://www.techscience.com/cmc/v74n3/50940}, ISSN = {1546-2226}, ABSTRACT = {The rise of social networking enables the development of multilingual Internet-accessible digital documents in several languages. The digital document needs to be evaluated physically through the Cross-Language Text Summarization (CLTS) involved in the disparate and generation of the source documents. Cross-language document processing is involved in the generation of documents from disparate language sources toward targeted documents. The digital documents need to be processed with the contextual semantic data with the decoding scheme. This paper presented a multilingual cross-language processing of the documents with the abstractive and summarising of the documents. The proposed model is represented as the Hidden Markov Model LSTM Reinforcement Learning (HMMlstmRL). First, the developed model uses the Hidden Markov model for the computation of keywords in the cross-language words for the clustering. In the second stage, bi-directional long-short-term memory networks are used for key word extraction in the cross-language process. Finally, the proposed HMMlstmRL uses the voting concept in reinforcement learning for the identification and extraction of the keywords. The performance of the proposed HMMlstmRL is 2% better than that of the conventional bi-direction LSTM model.}, DOI = {10.32604/cmc.2023.034072} }