sentences with the terms being fine-tuned and without
them. This will make it possible to better evaluate
when the fine-tuning should be stopped (the models
have learnt the new terms and other terminology is
not damaged). The larger validation set will provide
better evaluation of the fine-tuning results.
It was also observed that unlike the METEOR
metric, the BLEU metric is very strict and often is
equal 0 even if the translation is correct. On the
contrary, the Meteor metric may not take into account
the correctness of particular terms since it applies
synonyms to the evaluation. As a result, it is
recommended to use the BLEU metric on a
reasonably large validation set for evaluation of early
stopping point during fine-tuning, and the Meteor
metric for the final evaluation of the translation
quality.
ACKNOWLEDGEMENTS
The paper is due to the collaboration between SPC
RAS and Festo SE & Co. KG. The methodology and
experiment setup (sec. 3) are partially due to the State
Research, project number FFZF-2022-0005.
REFERENCES
Chen, Ben et al. 2021. “Transformer-Based Language
Model Fine-Tuning Methods for COVID-19 Fake
News Detection.” Communications in Computer and
Information Science 1402: 83–92.
Cioffi, Raffaele et al. 2020. “Artificial Intelligence and
Machine Learning Applications in Smart Production:
Progress, Trends, and Directions.” Sustainability 12(2):
492.
Denkowski, Michael, and Alon Lavie. 2014. “Meteor
Universal: Language Specific Translation Evaluation
for Any Target Language.” In Proceedings of the Ninth
Workshop on Statistical Machine Translation,
Stroudsburg, PA, USA: Association for Computational
Linguistics, 376–80.
Devlin, Jacob, Ming-Wei Chang, Kenton Lee, and Kristina
Toutanova. 2019. “BERT: Pre-Training of Deep
Bidirectional Transformers for Language
Understanding.” In Proceedings of NAACL-HLT 2019,
4171–86.
Fan, Angela et al. 2021. “Beyond English-Centric
Multilingual Machine Translation.” Journal of
Machine Learning Research 22: 1–48..
Google. 2022. “Google Tranlate.”
Jooste, Wandri, Rejwanul Haque, and Andy Way. 2021.
“Philipp Koehn: Neural Machine Translation.”
Machine Translation 35(2): 289–99.
Liu, Yinhan et al. 2020. “Multilingual Denoising Pre-
Training for Neural Machine Translation.”
Transactions of the Association for Computational
Linguistics 8: 726–42..
Mahata, Sainik Kumar, Dipankar Das, and Sivaji
Bandyopadhyay. 2019. “MTIL2017: Machine
Translation Using Recurrent Neural Network on
Statistical Machine Translation.” Journal of Intelligent
Systems 28(3): 447–53.
Microsoft. 2022. “Microsoft Bing Translator.”
Mishra, Sudhanshu, Shivangi Prasad, and Shubhanshu
Mishra. 2020. “Multilingual Joint Fine-Tuning of
Transformer Models for Identifying Trolling,
Aggression and Cyberbullying at TRAC 2020.” In
Proceedings of the Second Workshop on Trolling,
Aggression and Cyberbullying at Language Resources
and Evaluation Conference (LREC 2020), European
Language Resources Association (ELRA), 120–25.
Papineni, Kishore, Salim Roukos, Todd Ward, and Wei-
Jing Zhu. 2002. “BLEU: A Method for Automatic
Evaluation of Machine Translation.” In Proceedings of
the 40th Annual Meeting of the Association for
Computational Linguistics (ACL), Philadelphia, 311–
18.
Stahlberg, Felix. 2020. “Neural Machine Translation: A
Review.” Journal of Artificial Intelligence Research
69: 343–418.
Tan, Zhixing et al. 2020. “Neural Machine Translation: A
Review of Methods, Resources, and Tools.” AI Open 1:
5–21.
Tiedemann, Jörg, and Santhosh Thottingal. 2020. “OPUS-
MT: Building Open Translation Services for the
World.” In Proceedings of the 22nd Annual Conference
of the European Association for Machine Translation,
479–80.
Usuga Cadavid, Juan Pablo et al. 2020. “Machine Learning
Applied in Production Planning and Control: A State-
of-the-Art in the Era of Industry 4.0.” Journal of
Intelligent Manufacturing 31(6): 1531–58.
Vathsala, M. K., and Ganga Holi. 2020. “RNN Based
Machine Translation and Transliteration for Twitter
Data.” International Journal of Speech Technology
23(3): 499–504.
Wang, Xu, Chunyang Chen, and Zhenchang Xing. 2019.
“Domain-Specific Machine Translation with Recurrent
Neural Network for Software Localization.” Empirical
Software Engineering 24(6): 3514–45.
Wolf, Thomas et al. 2020. “HuggingFace’s Transformers:
State-of-the-Art Natural Language Processing.” In
Proceedings of the 2020 Conference on Empirical
Methods in Natural Language Processing: System
Demonstrations, Stroudsburg, PA, USA, ACL, 38–45.
Zhang, JiaJun, and ChengQing Zong. 2020. “Neural
Machine Translation: Challenges, Progress and
Future.” Science China Technological Sciences 63(10):
2028–50.