A Reverse Positional Encoding Multi-Head Attention-Based Neural Machine Translation Model for Arabic Dialectsopen access
- Authors
- Baniata, Laith H.; Kang, Sangwoo; Ampomah, Isaac K. E.
- Issue Date
- Oct-2022
- Publisher
- MDPI
- Keywords
- reverse positional encoding (RPE); multi-head attention; neural machine translation (NMT); Arabic dialects; MSA
- Citation
- MATHEMATICS, v.10, no.19
- Journal Title
- MATHEMATICS
- Volume
- 10
- Number
- 19
- URI
- https://scholarworks.bwise.kr/gachon/handle/2020.sw.gachon/85855
- DOI
- 10.3390/math10193666
- ISSN
- 2227-7390
- Abstract
- Languages with a grammatical structure that have a free order for words, such as Arabic dialects, are considered a challenge for neural machine translation (NMT) models because of the attached suffixes, affixes, and out-of-vocabulary words. This paper presents a new reverse positional encoding mechanism for a multi-head attention (MHA) neural machine translation (MT) model to translate from right-to-left texts such as Arabic dialects (ADs) to modern standard Arabic (MSA). The proposed model depends on an MHA mechanism that has been suggested recently. The utilization of the new reverse positional encoding (RPE) mechanism and the use of sub-word units as an input to the self-attention layer improve this sublayer for the proposed model's encoder by capturing all dependencies between the words in right-to-left texts, such as AD input sentences. Experiments were conducted on Maghrebi Arabic to MSA, Levantine Arabic to MSA, Nile Basin Arabic to MSA, Gulf Arabic to MSA, and Iraqi Arabic to MSA. Experimental analysis proved that the proposed reverse positional encoding MHA NMT model was efficiently able to handle the open grammatical structure issue of Arabic dialect sentences, and the proposed RPE MHA NMT model enhanced the translation quality for right-to-left texts such as Arabic dialects.
- Files in This Item
- There are no files associated with this item.
- Appears in
Collections - IT융합대학 > 소프트웨어학과 > 1. Journal Articles
![qrcode](https://api.qrserver.com/v1/create-qr-code/?size=55x55&data=https://scholarworks.bwise.kr/gachon/handle/2020.sw.gachon/85855)
Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.