Optimizing Indonesian-Sundanese Bilingual Translation with Adam-Based Neural Machine Translation
This research seeks to construct an automatic translation between Indonesian and Sundanese languages based on the Neural Machine Translation (NMT) method. The model used in this study is the Long Short-Term Memory (LSTM) type, which carries out an encoder-decoder structure model learned with Bible d...
Saved in:
Main Authors: | , , , , |
---|---|
Format: | Article |
Language: | English |
Published: |
Ikatan Ahli Informatika Indonesia
2024-12-01
|
Series: | Jurnal RESTI (Rekayasa Sistem dan Teknologi Informasi) |
Subjects: | |
Online Access: | https://jurnal.iaii.or.id/index.php/RESTI/article/view/6116 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Summary: | This research seeks to construct an automatic translation between Indonesian and Sundanese languages based on the Neural Machine Translation (NMT) method. The model used in this study is the Long Short-Term Memory (LSTM) type, which carries out an encoder-decoder structure model learned with Bible data. The text translation here was conducted in different epochs to optimize the process, followed by the Adam optimization algorithm. Testing the Adam optimizer with different epoch settings yields a BLEU score for Indonesian to Sundanese translations of 0.991785, higher than the performance of the None optimizer. Experimental results demonstrate that Indonesian to Sundanese translation using Adam optimization with 1000 epochs consistently performed better in BLEU - Bilingual Evaluation Understudy - scoring than Sundanese to Indonesian translation. Limitations of the research were also put forth, particularly technical issues related to the collection of data and the Sundanese language’s complex grammatical features, that the model can only partially express, honorifics, and the problem of polysemy. Also, it must be mentioned that no special hyperparameter selection was performed, as parameters were chosen randomly. In future studies, transformer-based models can be investigated since these architectures will better deal with complex language via their self-attention mechanism. |
---|---|
ISSN: | 2580-0760 |