Download PDFOpen PDF in browser

Enhancing English-Persian Neural Machine Translation with a Large-Scale Parallel Dataset and Relative Position Representations

EasyChair Preprint 15545

6 pagesDate: December 9, 2024

Abstract

Transformer-based models have revolutionized neural machine translation (NMT), particularly with the introduction of the encoder-decoder architecture. However, training these models effectively often requires large amounts of parallel data or pre-training on massive unlabeled corpora. In the context of English-Persian translation, the lack of extensive parallel datasets has hindered progress. To address this, we introduce a new dataset of 4 million English-Persian parallel sentences that span various topics. Without any pre-training on unlabeled data, our model achieves a BLEU score of 47 on the PEPC benchmark and 35 on the MIZAN benchmark, demonstrating strong performance. We used Transformers with relative position representations, enabling the model to generalize to sequence lengths not seen during training. To promote further research and reproducibility, we have open-sourced both the dataset and the trained model, supporting advancements in English-Persian NMT.

Keyphrases: Attention Mechanism, BLEU score, English-Persian dataset, Neural Machine Translation, transformer

BibTeX entry
BibTeX does not have the right entry for preprints. This is a hack for producing the correct reference:
@booklet{EasyChair:15545,
  author    = {Alireza Kamyab and Negar Baghaei Nejad and Alireza Akhavanpour},
  title     = {Enhancing English-Persian Neural Machine Translation with a Large-Scale Parallel Dataset and Relative Position Representations},
  howpublished = {EasyChair Preprint 15545},
  year      = {EasyChair, 2024}}
Download PDFOpen PDF in browser