Optimizing Indonesian-Sundanese Bilingual Translation with Adam-Based Neural Machine Translation

This research seeks to construct an automatic translation between Indonesian and Sundanese languages based on the Neural Machine Translation (NMT) method. The model used in this study is the Long Short-Term Memory (LSTM) type, which carries out an encoder-decoder structure model learned with Bible d...

Full description

Saved in:
Bibliographic Details
Main Authors: Anita Qotrun Nada, Aji Prasetya Wibawa, Dhea Fanny Putri Syarifa, Erliana Fajarwati, Fadia Irsania Putri
Format: Article
Language:English
Published: Ikatan Ahli Informatika Indonesia 2024-12-01
Series:Jurnal RESTI (Rekayasa Sistem dan Teknologi Informasi)
Subjects:
Online Access:https://jurnal.iaii.or.id/index.php/RESTI/article/view/6116
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:This research seeks to construct an automatic translation between Indonesian and Sundanese languages based on the Neural Machine Translation (NMT) method. The model used in this study is the Long Short-Term Memory (LSTM) type, which carries out an encoder-decoder structure model learned with Bible data. The text translation here was conducted in different epochs to optimize the process, followed by the Adam optimization algorithm. Testing the Adam optimizer with different epoch settings yields a BLEU score for Indonesian to Sundanese translations of 0.991785, higher than the performance of the None optimizer. Experimental results demonstrate that Indonesian to Sundanese translation using Adam optimization with 1000 epochs consistently performed better in BLEU - Bilingual Evaluation Understudy - scoring than Sundanese to Indonesian translation. Limitations of the research were also put forth, particularly technical issues related to the collection of data and the Sundanese language’s complex grammatical features, that the model can only partially express, honorifics, and the problem of polysemy. Also, it must be mentioned that no special hyperparameter selection was performed, as parameters were chosen randomly. In future studies, transformer-based models can be investigated since these architectures will better deal with complex language via their self-attention mechanism.
ISSN:2580-0760