Skip to main content

ORIGINAL RESEARCH article

Front. Hum. Neurosci.
Sec. Speech and Language
Volume 19 - 2025 | doi: 10.3389/fnhum.2025.1498297
This article is part of the Research Topic Emerging Techniques in Arabic Natural Language Processing View all articles

Advancing Arabic Dialect Detection with Hybrid Stacked Transformer Models

Provisionally accepted
  • 1 Faculty of Computing and Artificial Intelligence, South Valley University, Hurghada, Egypt
  • 2 Department of Computer Science, College of Computer and Information Sciences, Imam Muhammad ibn Saud Islamic University, Riyadh, Saudi Arabia
  • 3 Faculty of Science, Aswan University, Aswan, Aswan, Egypt
  • 4 College of Computer and Information Sciences, Imam Muhammad ibn Saud Islamic University, Riyadh, Saudi Arabia
  • 5 Department of Computer Science and Engineering, College of Informatics, Korea University, Seoul, Republic of Korea

The final, formatted version of the article will be published soon.

    The rapid expansion of dialectally unique Arabic material on social media and the internet highlights how important it is to categorize dialects accurately to maximize a variety of Natural Language Processing (NLP) applications. The improvement in classification performance highlights the wider variety of linguistic variables that the model can capture, providing a reliable solution for precise Arabic dialect recognition and improving the efficacy of NLP applications. Recent advances in deep learning (DL) models have shown promise in overcoming potential challenges in identifying Arabic dialects. In this paper, we propose a novel stacking model based on two transformer models, i.e., Bert-Base-Arabertv02 and Dialectal-Arabic-XLM-R-Base, to enhance the classification of dialectal Arabic. The proposed model consists of two levels, including base models and meta-learners. In the proposed model, Level 1 generates class probabilities from two transformer models for training and testing sets, which are then used in Level 2 to 1 Sample et al.

    Keywords: Arabic dialects, Bert-Base-Arabertv02, Dialectal-Arabic-XLM-R-Base, transformer, Knowledge representation, nlp, deep learning, Stacking model

    Received: 18 Sep 2024; Accepted: 24 Jan 2025.

    Copyright: © 2025 Saleh, Almohimeed, Hassan, Ibrahim, Alsamhi, Hassan and Mostafa. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

    * Correspondence: Hager Saleh, Faculty of Computing and Artificial Intelligence, South Valley University, Hurghada, Egypt

    Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.