Seamless Textual Version Using with MarianMT Technique

Authors

  • B. Snehalatha M.C.A Student, Department of M.C.A, KMMIPS, Tirupati (D.t) , Andhra Pradesh, India Author
  • S. Noortaj Assistant Professor, Department of M.C.A, KMMIPS, Tirupati (D.t), Andhra Pradesh, India Author

Keywords:

Context-aware translation, MarianMT, BERT, Indian languages, Hugging Face, speech recognition, text-to-speech, adaptive learning, natural language processing, language acquisition

Abstract

This project presents a context-aware, multilingual translation system that enhances the accuracy and fluidity of translations for multiple Indian languages. The system combines MarianMT, a robust machine translation model, with BERT for contextual understanding, enabling more accurate translations by capturing the nuances of word relationships within each sentence. Fine-tuning the MarianMT model with an English-Hindi parallel corpus further improves the model’s sensitivity to linguistic subtleties, idiomatic expressions, and cultural references unique to Hindi. Efficiency is optimized through mixed-precision training and gradient accumulation, allowing the model to handle large datasets effectively while minimizing computational overhead. To extend functionality across Indian languages, the system incorporates models from the HelsinkiNLP OPUSMT series, accessed via the Hugging Face transformers library. This integration supports real-time translation for Hindi, Marathi, Telugu, Kannada, Tamil, Bengali, and Gujarati, bridging language barriers and enhancing communication. The system also includes speech-to-text and text-to-speech capabilities, powered by libraries like speech_recognition and gTTS, enabling seamless conversion between spoken and written language. An adaptive learning component is introduced, utilizing machine learning algorithms to generate personalized quizzes based on user interaction and performance, promoting effective language learning. By combining advanced natural language processing with interactive educational tools, this translation system serves both as a robust language translation solution and as an innovative platform for language acquisition, applicable in educational and cross-cultural communication contexts.

Downloads

Download data is not yet available.

References

A. Vaswani et al., "Attention is all you need," vol. 30, 2017.

A. Radford et al., "Learning transferable visual models from natural language supervision," in International conference on machine learning, 2021, pp. 8748-8763: PmLR.

D. Bahdanau, K. Cho, and Y. J. a. p. a. Bengio, "Neural machine translation by jointly learning to align and translate," 2014.

A. Graves, A.-r. Mohamed, and G. Hinton, "Speech recognition with deep recurrent neural networks," in 2013 IEEE international conference on acoustics, speech and signal processing, 2013, pp. 6645-6649: Ieee.

J. Devlin, M.-W. Chang, K. Lee, and K. Toutanova, "Bert: Pre-training of deep bidirectional transformers for language understanding," in Proceedings of the 2019 conference of the North American chapter of the association for computational linguistics: human language technologies, volume 1 (long and short papers), 2019, pp. 4171-4186.

Downloads

Published

30-05-2025

Issue

Section

Research Articles

How to Cite

[1]
B. Snehalatha and S. Noortaj, “Seamless Textual Version Using with MarianMT Technique ”, Int J Sci Res Sci Eng Technol, vol. 12, no. 3, pp. 542–550, May 2025, Accessed: Jun. 03, 2025. [Online]. Available: https://ijsrset.com/index.php/home/article/view/IJSRSET251278

Similar Articles

1-10 of 255

You may also start an advanced similarity search for this article.