Home / Advanced Search

  • Title/Keywords

  • Author/Affliations

  • Journal

  • Article Type

  • Start Year

  • End Year

Update SearchingClear
  • Articles
  • Online
Search Results (2)
  • Open Access

    ARTICLE

    Multi-Task Learning Model with Data Augmentation for Arabic Aspect-Based Sentiment Analysis

    Arwa Saif Fadel1,2,*, Osama Ahmed Abulnaja1, Mostafa Elsayed Saleh1

    CMC-Computers, Materials & Continua, Vol.75, No.2, pp. 4419-4444, 2023, DOI:10.32604/cmc.2023.037112

    Abstract Aspect-based sentiment analysis (ABSA) is a fine-grained process. Its fundamental subtasks are aspect term extraction (ATE) and aspect polarity classification (APC), and these subtasks are dependent and closely related. However, most existing works on Arabic ABSA content separately address them, assume that aspect terms are preidentified, or use a pipeline model. Pipeline solutions design different models for each task, and the output from the ATE model is used as the input to the APC model, which may result in error propagation among different steps because APC is affected by ATE error. These methods are impractical for real-world scenarios where the… More >

  • Open Access

    ARTICLE

    Embedding Extraction for Arabic Text Using the AraBERT Model

    Amira Hamed Abo-Elghit1,*, Taher Hamza1, Aya Al-Zoghby2

    CMC-Computers, Materials & Continua, Vol.72, No.1, pp. 1967-1994, 2022, DOI:10.32604/cmc.2022.025353

    Abstract Nowadays, we can use the multi-task learning approach to train a machine-learning algorithm to learn multiple related tasks instead of training it to solve a single task. In this work, we propose an algorithm for estimating textual similarity scores and then use these scores in multiple tasks such as text ranking, essay grading, and question answering systems. We used several vectorization schemes to represent the Arabic texts in the SemEval2017-task3-subtask-D dataset. The used schemes include lexical-based similarity features, frequency-based features, and pre-trained model-based features. Also, we used contextual-based embedding models such as Arabic Bidirectional Encoder Representations from Transformers (AraBERT). We… More >

Displaying 1-10 on page 1 of 2. Per Page