Japanese waka translation supported by internet of things and artificial intelligence technology

Abstract With the advancement of internet of things (IoT) and artificial intelligence (AI) technology, access to large-scale bilingual parallel data has become more efficient, thereby accelerating the development and application of machine translation. Given the increasing cultural exchanges between...

Full description

Saved in:
Bibliographic Details
Main Author: Rizhong Shen
Format: Article
Language:English
Published: Nature Portfolio 2025-01-01
Series:Scientific Reports
Subjects:
Online Access:https://doi.org/10.1038/s41598-025-85184-y
Tags: Add Tag
No Tags, Be the first to tag this record!
_version_ 1841544606771052544
author Rizhong Shen
author_facet Rizhong Shen
author_sort Rizhong Shen
collection DOAJ
description Abstract With the advancement of internet of things (IoT) and artificial intelligence (AI) technology, access to large-scale bilingual parallel data has become more efficient, thereby accelerating the development and application of machine translation. Given the increasing cultural exchanges between China and Japan, many scholars have begun to study the Chinese translation of Japanese waka poetry. Based on this, the study first explores the structure of waka and the current state of its Chinese translations, analyzing existing translation disputes and introducing a data collection method for waka using IoT. Then, an optimized neural machine translation model is proposed, which integrates a Bidirectional Long Short-Term Memory (Bi-LSTM) network, vertical Tree-LSTM, and an attention mechanism into the Transformer framework. Experimental results demonstrate that the three optimized models—Transformer + Bi-LSTM, Transformer + Tree-LSTM, and Transformer + Tree-LSTM + Attention—outperform the baseline Transformer model on both public and waka datasets. The BLEU scores of the models on the public dataset were 23.71, 23.95, and 24.12, respectively. Notably, on the waka dataset, the Transformer + Tree-LSTM + Attention model achieved the highest BLEU score of 20.65, demonstrating a significant advantage in capturing waka’s unique features and contextual information. This study offers new methods to enhance the quality of Chinese-Japanese translation, promoting cultural exchange and understanding.
format Article
id doaj-art-9b50cd92ea3c49aba9c78bc5fe11afec
institution Kabale University
issn 2045-2322
language English
publishDate 2025-01-01
publisher Nature Portfolio
record_format Article
series Scientific Reports
spelling doaj-art-9b50cd92ea3c49aba9c78bc5fe11afec2025-01-12T12:24:01ZengNature PortfolioScientific Reports2045-23222025-01-0115111410.1038/s41598-025-85184-yJapanese waka translation supported by internet of things and artificial intelligence technologyRizhong Shen0School of Foreign Languages, Quanzhou Normal UniversityAbstract With the advancement of internet of things (IoT) and artificial intelligence (AI) technology, access to large-scale bilingual parallel data has become more efficient, thereby accelerating the development and application of machine translation. Given the increasing cultural exchanges between China and Japan, many scholars have begun to study the Chinese translation of Japanese waka poetry. Based on this, the study first explores the structure of waka and the current state of its Chinese translations, analyzing existing translation disputes and introducing a data collection method for waka using IoT. Then, an optimized neural machine translation model is proposed, which integrates a Bidirectional Long Short-Term Memory (Bi-LSTM) network, vertical Tree-LSTM, and an attention mechanism into the Transformer framework. Experimental results demonstrate that the three optimized models—Transformer + Bi-LSTM, Transformer + Tree-LSTM, and Transformer + Tree-LSTM + Attention—outperform the baseline Transformer model on both public and waka datasets. The BLEU scores of the models on the public dataset were 23.71, 23.95, and 24.12, respectively. Notably, on the waka dataset, the Transformer + Tree-LSTM + Attention model achieved the highest BLEU score of 20.65, demonstrating a significant advantage in capturing waka’s unique features and contextual information. This study offers new methods to enhance the quality of Chinese-Japanese translation, promoting cultural exchange and understanding.https://doi.org/10.1038/s41598-025-85184-yJapanese wakaInternet of thingsNeural machine translationTransformerLong short-term memory network
spellingShingle Rizhong Shen
Japanese waka translation supported by internet of things and artificial intelligence technology
Scientific Reports
Japanese waka
Internet of things
Neural machine translation
Transformer
Long short-term memory network
title Japanese waka translation supported by internet of things and artificial intelligence technology
title_full Japanese waka translation supported by internet of things and artificial intelligence technology
title_fullStr Japanese waka translation supported by internet of things and artificial intelligence technology
title_full_unstemmed Japanese waka translation supported by internet of things and artificial intelligence technology
title_short Japanese waka translation supported by internet of things and artificial intelligence technology
title_sort japanese waka translation supported by internet of things and artificial intelligence technology
topic Japanese waka
Internet of things
Neural machine translation
Transformer
Long short-term memory network
url https://doi.org/10.1038/s41598-025-85184-y
work_keys_str_mv AT rizhongshen japanesewakatranslationsupportedbyinternetofthingsandartificialintelligencetechnology