Joint NMT models for text conversion between traditional Mongolian script and cyrillic Mongolian: a comparative study

Authors

  • Uuganbaatar Dulamragchaa Institute of Mathematics and Digital Technology, Mongolian Academy of Sciences, Mongolia https://orcid.org/0000-0001-7392-638X
  • Nomuundalai Batbileg Institute of Mathematics and Digital Technology, Mongolian Academy of Sciences, Ulaanbaatar 13330, Mongolia https://orcid.org/0009-0005-7259-4068
  • Sainbaya Batbaatar Institute of Mathematics and Digital Technology, Mongolian Academy of Sciences, Ulaanbaatar 13330, Mongolia
  • Purevsuren Tumurbaatar Institute of Language and Literature, Mongolian Academy of Sciences, Ulaanbaatar 13330, Mongolia https://orcid.org/0000-0002-6657-9794
  • Bilguutei Narantuya Institute of Mathematics and Digital Technology, Mongolian Academy of Sciences, Ulaanbaatar 13330, Mongolia https://orcid.org/0009-0003-3957-8344
  • Baatar Nyamkhuu Institute of Mathematics and Digital Technology, Mongolian Academy of Sciences, Ulaanbaatar 13330, Mongolia https://orcid.org/0009-0009-8292-8569
  • Tungalagtamir Bold Institute of Mathematics and Digital Technology, Mongolian Academy of Sciences, Ulaanbaatar 13330, Mongolia https://orcid.org/0009-0003-1739-9425
  • Bayarchimeg Enkhtaivan Institute of Language and Literature, Mongolian Academy of Sciences, Ulaanbaatar 13330, Mongolia

DOI:

https://doi.org/10.5564/jimdt.v6i1.3758

Keywords:

Machine Translation, Transformer, Neural machine translation (NMT), Deep Learning

Abstract

The research aims to develop a high-precision machine learning model for translating between Mongolian and Cyrillic scripts. Although Mongolian and Cyrillic scripts are different scripts of the Mongolian language, their sentence structure, number of words in a sentence, and position are the same. This article presents the results of our research on neural machine translation models tested for translating between the two scripts. The Seq2Seq and Transformer models with attention mechanisms were trained using each tokenization method at the character, subword, and word levels. The accomplishments of these models were assessed regarding the conversion quality, computational efficiency, and the ability to handle the unique characteristics of the two scripts, and the benefits and drawbacks of each model were summarized. It is believed that this comparison helps select the most suitable NMT model for similar tasks. The developed model has a vast potential for development and application as a system for translating between scripts, and it will facilitate some clerical work in public and private sector organizations at all levels providing services in Cyrillic and Mongolian scripts.

Downloads

Download data is not yet available.
Abstract
146
PDF
113

References

[1] I. Sutskever, O. Vinyals, and Q. V. Le, “Sequence-to-Sequence Learning with Neural Networks,” Advances in Neural Information Processing Systems, Vol. 27, Curran Associates Inc., 2014.

[2] Proceedings of the 27th International Conference on Neural Information Processing Systems, Curran Associates Inc., 2014.

[3] D. Bahdanau, K. Cho, and Y. Bengio, “Neural Machine Translation by Jointly Learning to Align and Translate,” arXiv preprint arXiv:1409.0473, 2015.

[4] A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. N. Gomez, L. Kaiser, and I. Polosukhin, “Attention Is All You Need,” Advances in Neural Information Processing Systems, edited by I. Guyon, U. V. Luxburg, S. Bengio, H. Wallach, R. Fergus, S. Vishwanathan, R. Garnett, and D. Precup, Curran Associates Inc., 2017, pp. 5998–6008.

[5] R. Sennrich, B. Haddow, and A. Birch, “Neural machine translation of rare words with subword units,” In Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics, pp. 1715–1725, 2016. https://doi.org/10.18653/v1/P16-1162

[6] A. Mao, et al., “Cross-Entropy Loss Functions: Theoretical Analysis and Applications,” In International Conference on Machine Learning, 2023, pp. 23803–23828.

Downloads

Published

2024-12-27

How to Cite

Dulamragchaa, U., Batbileg, N., Batbaatar, S., Tumurbaatar, P., Narantuya, B., Nyamkhuu, B., … Enkhtaivan, B. (2024). Joint NMT models for text conversion between traditional Mongolian script and cyrillic Mongolian: a comparative study. Journal of Institute of Mathematics and Digital Technology, 6(1), 99–108. https://doi.org/10.5564/jimdt.v6i1.3758

Issue

Section

Articles