
A Comparative Study of Machine Translation Tools for English-to-Marathi Language Translation | IJCT Volume 12 – Issue 5 | IJCT-V12I5P62

International Journal of Computer Techniques
ISSN 2394-2231
Volume 12, Issue 5 | Published: September β October 2025
Author
Taiba Imtiyaz Tamboli , Simran Mohan Jadhav , Rubina Sheikh
Table of Contents
ToggleAbstract
The rapid advancement of artificial intelligence has led to the development of highly efficient machine translation systems such as Google Translate, ChatGPT and Gemini. This study presents a comparative evaluation of three machine translation systems β Google Translate, ChatGPT and Gemini β based on both manual and automated assessment methods. Manual evaluation focused on accuracy, fluency, and adequacy, while automated evaluation was conducted using the BLEU (Bilingual Evaluation Understudy) score, calculated through Python programming. A One-Way ANOVA test revealed a statistically significant difference among the BLEU scores (F(2,147) = 10.26109, p < 0.05), indicating that translation quality varied across systems. Tukeyβs post-hoc test showed that Gemini performed significantly better than both Google and ChatGPT, while no significant difference was observed between Google and ChatGPT. Error analysis further supported these results, with Gemini showing minimal grammatical errors, Google displaying moderate lexical and idiomatic issues, and ChatGPT exhibiting varied misformation and semantic errors. Overall, Gemini demonstrated the highest translation accuracy and fluency, validating its superior performance among the three systems.
Keywords
Machine Translation, English-Marathi, ChatGPT, Gemini, Google Translate, Error AnalysisConclusion
This paper presented a comprehensive comparative analysis of Google Translate, ChatGPT, and Gemini for English-to-Marathi machine translation using both manual and automated evaluation methods. Manual assessment based on accuracy, fluency, and adequacy revealed that Gemini consistently produced more natural and contextually appropriate translations, followed by ChatGPT and Google Translate.
Automated evaluation using the BLEU score, computed through Python, supported these findings. The one-way ANOVA test yielded an F-value of 10.26109, indicating statistically significant differences among the mean BLEU scores of the three systems. Subsequent Tukeyβs post-hoc pairwise analysis showed that while the difference between Google and ChatGPT was not significant, both Google vs. Gemini and ChatGPT vs. Gemini comparisons were statistically significant. This confirms Geminiβs superior translation quality in terms of automated evaluation metrics.
The error analysis further revealed that Gemini produced the fewest linguistic and semantic errors, while ChatGPT exhibited frequent lexical and grammatical misformations. Google Translate demonstrated balanced performance but occasionally struggled with idiomatic and context-dependent phrases.
Overall, the findings highlight that Gemini outperforms Google Translate and ChatGPT in both manual and automated assessments for English-to-Marathi translation. The study emphasizes the growing reliability of advanced AI-driven systems like Gemini and ChatGPT in handling low-resource regional languages, while also identifying areas for future improvement in context understanding and idiomatic accuracy.
References
[1]M. Aiken and Sh. Balan, βAn analysis of Google Translate accuracy,β Translation Journal, vol. 16, no. 2, 2011.
[2]H. Ghasemi and M. Hashemian, βA comparative study of Google Translate translations: An error analysis of English-to-Persian and Persian-to-English translations,β English Language Teaching, vol. 9, no. 3, pp. 13β17, 2016.
[3]K. Papineni, S. Roukos, T. Ward, and W. J. Zhu, βBLEU: A method for automatic evaluation of machine translation,β in Proc. 40th Annual Meeting on Association for Computational Linguistics (ACL), 2002, pp. 311β318.
[4]R. Sennrich, B. Haddow, and A. Birch, βNeural machine translation of rare words with subword units,β in Proc. 54th Annual Meeting of the Association for Computational Linguistics (ACL), 2016, pp. 1715β1725.
[5]P. Goyal and G. S. Lehal, βHindiβMarathi machine translation system using rule-based approach,β Journal of Computing and Information Technology, vol. 16, no. 3, pp. 225β232, 2008.
[6]M. Joshi, S. Singh, and S. Saini, βComparative study of machine translation tools for Indian languages,β Int. J. Adv. Comput. Sci. Appl. (IJACSA), vol. 12, no. 3, pp. 45β52, 2021.
[7]M. PopoviΔ, βA comparative analysis of evaluation metrics for machine translation,β Language Resources and Evaluation, vol. 49, no. 3, pp. 679β705, 2015.
[8]E. Loffler, βError typologies for machine translation quality assessment,β Linguistica Antverpiensia, vol. 20, pp. 230β245, 2021.
[9]P. Rao, L. M. McGee, and C. A. Seideman, βA comparative assessment of ChatGPT vs. Google Translate for the translation of patient instructions,β J. Med. Artif. Intell., vol. 6, no. 11, 2023.
[10]M. M. Obeidat, A. S. Haider, S. Abu Tair, and Y. Sahari, βAnalyzing the performance of Gemini, ChatGPT, and Google Translate in rendering English idioms into Arabic,β FWU J. Soc. Sci., vol. 18, no. 4, pp. 1β18, 2024.
[11]R. Ranathunga, R. Sennrich, et al., βOverview and challenges of machine translation for contextually-rich, low-resource languages,β PMC Survey, 2023.
[12]B. Haddow, R. Bawden, A. V. Miceli Barone, J. Helcl, and A. Birch, βSurvey of low-resource machine translation,β ACL Anthology, 2022.
Journal Covers
IJCT Important Links
Β© 2025 International Journal of Computer Techniques (IJCT).









