Hierarchical Triple Model of Hybrid Neural Machine Translation

Bat-Erdene Batsukh


For more than a decade, PMT and SMT models have dominated the field of machine translation, and neural machine translation has emerged as a new paradigm for machine translation by the 2015. Neural machine translation provides a simple modeling mechanism that is easy to use in practice and science. Thus, it does not require concepts such as word ranking, a key component of the statistical machine translation. While this simplicity may be seen as an advantage, on the other hand, the lack of careful spelling is to lose control of the translation. Even tough, the neural machine translation is more flexible in terms of translations that don’t exactly match the training data. This provides more opportunities for such models, but exempts translation from pre-determined restrictions. Failure to connect specific words can make it difficult to connect the target words you create to the original word. The widespread use of neural machine translation system has the advantage of allowing users to translate certain terms and translate uneducated data to a certain extent. In some cases, however, the structure and the grammar boundary of a sentence is often distorted. The paper is intended to address issues such as the control of neural machine translation, more accurate translation of unidentified data, the accuracy of sentence structure and grammar boundaries. To solve this problem, modern translation theory led to the hybrid model of machine translation. Our model is expansion of this hybrid model with a sentence and a grammar boundary. We named this model as hierarchical triple model (HTM).


English-Mongolian translation, Hybrid model expansion, Grammar boundary, Sentence structure Hierarchical triple model

Full Text:



Alkhouli, T., Bretschner, G., Peter, J.-T., Hethnawi, M., Guta, A., & Ney, H. (2016). Alignment-Based Neural Machine Translation. Proceedings of the First Conference on Machine Translation: Volume 1, Research Papers, 54–65. https://doi.org/10.18653/v1/W16-2206

Brown, P. F., Cocke, J., Della Pietra, S. A., Della Pietra, V. J., Jelinek, F., Lafferty, J. D., Mercer, R. L., & Roossin, P. S. (1990). A statistical approach to machine translation. Computational Linguistics, 79–85. https://aclanthology.org/J90-2002/

Brown, P. F., Della Pietra, S. A., Della Pietra, V. J., & Mercer, R. L. (1993). The Mathematics of Statistical Machine Translation: Parameter Estimation. Computational Linguistics, 19(2), 263–311. https://aclanthology.org/J93-2003

Dozat, T., Qi, P., & Manning, C. D. (2017). Stanford’s Graph-based Neural Dependency Parser at the CoNLL 2017 Shared Task. Proceedings of the CoNLL 2017 Shared Task: Multilingual Parsing from Raw Text to Universal Dependencies, 20–30. https://doi.org/10.18653/v1/K17-3002

Kalchbrenner, N., & Blunsom, P. (2013). Recurrent Continuous Translation Models. Proceedings of the 2013 Conference on Empirical Methods in Natural Language Processing, 1700–1709. https://aclanthology.org/D13-1176

Koehn, P., Och, F. J., & Marcu, D. (2003). Statistical Phrase-Based Translation. Proceedings of the 2003 Human Language Technology Conference of the North American Chapter of the Association for Computational Linguistics, 127–133. https://aclanthology.org/N03-1017

Lison, P., & Tiedemann, J. (2016). OpenSubtitles2016: Extracting Large Parallel Corpora from Movie and TV Subtitles. Proceedings of the 10th International Conference on Language Resources and Evaluation (LREC 2016). http://www.opensubtitles.org.

Och, F. J. (2003). Minimum Error Rate Training in Statistical Machine Translation. Proceedings of the 41st Annual Meeting of the Association for Computational Linguistics, 160–167. https://doi.org/10.3115/1075096.1075117

Och, F. J., & Ney, H. (2000). Improved Statistical Alignment Models. Proceedings of the 38th Annual Meeting of the Association for Computational Linguistics, 440–447. https://doi.org/10.3115/1075218.1075274

Papineni, K., Roukos, S., Ward, T., & Zhu, W.-J. (2002). Bleu: a Method for Automatic Evaluation of Machine Translation. Proceedings of the 40th Annual Meeting of the Association for Computational Linguistics, 311–318. https://doi.org/10.3115/1073083.1073135

Schwenk, H., Chaudhary, V., Sun, S., Gong, H., & Guzmán, F. (2019). WikiMatrix: Mining 135M Parallel Sentences in 1620 Language Pairs from Wikipedia. CoRR, abs/1907.05791. http://arxiv.org/abs/1907.05791

Snover, M., Dorr, B., Schwartz, R., Micciulla, L., & Makhoul, J. (2006). A Study of Translation Edit Rate with Targeted Human Annotation. Proceedings of the 7th Conference of the Association for Machine Translation in the Americas: Technical Papers, 223–231. https://aclanthology.org/2006.amta-papers.25

Tan, Z., Wang, S., Yang, Z., Chen, G., Huang, X., Sun, M., & Liu, Y. (2020). Neural machine translation: A review of methods, resources, and tools. AI Open, 1, 5–21. https://doi.org/10.1016/j.aiopen.2020.11.001

Vauquois, B. (1968). A survey of formal grammars and algorithms for recognition and transformation in mechanical translation. IFIP Congress. https://dblp.org/rec/conf/ifip/Vauquois68

Vogel, S., Ney, H., & Tillmann, C. (1996). HMM-Based Word Alignment in Statistical Translation. International Conference on Computational Linguistics, 836–841. https://aclanthology.org/C96-2141.pdf

Wang, H., & Huang, Y. (2003). Bondec-A Sentence Boundary Detector. Stanford, 1-9. https://nlp.stanford.edu/courses/cs224n/2003/fp/huangy/final_project.doc

Wang, W., Alkhouli, T., Zhu, D., & Ney, H. (2017). Hybrid Neural Network Alignment and Lexicon Model in Direct HMM for Statistical Machine Translation. Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), 125–131. https://doi.org/10.18653/v1/P17-2020

Zens, R., & Ney, H. (2008). Improvements in Dynamic Programming Beam Search for Phrase-based Statistical Machine Translation. International Workshop on Spoken Language Translation, 195–205. https://aclanthology.org/www.mt-archive.info/05/IWSLT-2008-Zens.pdf

Ziemski, M., Junczys-Dowmunt, M., & Pouliquen, B. (2016). The United Nations Parallel Corpus v1.0. Proceedings of the Tenth International Conference on Language Resources and Evaluation (LREC’16), 3530–3534. https://aclanthology.org/L16-1561

DOI: https://doi.org/10.7575/aiac.ijalel.v.11n.2p.38


  • There are currently no refbacks.

Creative Commons License
This work is licensed under a Creative Commons Attribution 4.0 International License.

2012-2022 (CC-BY) Australian International Academic Centre PTY.LTD

International Journal of Applied Linguistics and English Literature

To make sure that you can receive messages from us, please add the journal emails into your e-mail 'safe list'. If you do not receive e-mail in your 'inbox', check your 'bulk mail' or 'junk mail' folders.