Performance of machine translators in translating French medical research abstracts to English: A comparative study of DeepL, Google Translate, and CUBBITT (2024)

  • Journal List
  • PLoS One
  • PMC10833527

As a library, NLM provides access to scientific literature. Inclusion in an NLM database does not imply endorsem*nt of, or agreement with, the contents by NLM or the National Institutes of Health.
Learn more: PMC Disclaimer | PMC Copyright Notice

Performance of machine translators in translating French medical research abstracts to English: A comparative study of DeepL, Google Translate, and CUBBITT (1)

Link to Publisher's site

PLoS One. 2024; 19(2): e0297183.

Published online 2024 Feb 1. doi:10.1371/journal.pone.0297183

PMCID: PMC10833527

PMID: 38300946

Paul Sebo, Conceptualization, Data curation, Formal analysis, Methodology, Writing – original draftPerformance of machine translators in translating French medical research abstracts to English: A comparative study of DeepL, Google Translate, and CUBBITT (2)1,* and Sylvain de Lucia, Conceptualization, Data curation, Formal analysis, Methodology2

Iftikhar Ahmed Khan, Editor

Author information Article notes Copyright and License information PMC Disclaimer

Associated Data

Data Availability Statement

Abstract

Background

Non-English speaking researchers may find it difficult to write articles in English and may be tempted to use machine translators (MTs) to facilitate their task. We compared the performance of DeepL, Google Translate, and CUBBITT for the translation of abstracts from French to English.

Methods

We selected ten abstracts published in 2021 in two high-impact bilingual medical journals (CMAJ and Canadian Family Physician) and used nine metrics of Recall-Oriented Understudy for Gisting Evaluation (ROUGE-1 recall/precision/F1-score, ROUGE-2 recall/precision/F1-score, and ROUGE-L recall/precision/F1-score) to evaluate the accuracy of the translation (scores ranging from zero to one [= maximum]). We also used the fluency score assigned by ten raters to evaluate the stylistic quality of the translation (ranging from ten [= incomprehensible] to fifty [= flawless English]). We used Kruskal-Wallis tests to compare the medians between the three MTs. For the human evaluation, we also examined the original English text.

Results

Differences in medians were not statistically significant for the nine metrics of ROUGE (medians: min-max = 0.5246–0.7392 for DeepL, 0.4634–0.7200 for Google Translate, 0.4815–0.7316 for CUBBITT, all p-values > 0.10). For the human evaluation, CUBBITT tended to score higher than DeepL, Google Translate, and the original English text (median = 43 for CUBBITT, vs. 39, 38, and 40, respectively, p-value = 0.003).

Conclusion

The three MTs performed similarly when tested with ROUGE, but CUBBITT was slightly better than the other two using human evaluation. Although we only included abstracts and did not evaluate the time required for post-editing, we believe that French-speaking researchers could use DeepL, Google Translate, or CUBBITT when writing articles in English.

Introduction

The dominance of English as the publishing language can penalize non-English speaking researchers seeking to share their work, as the stylistic quality of articles can have an impact on their likelihood of being published and/or cited by other publications [1]. In order to improve their chances of publication in English-language journals, researchers often rely on professional translation services to improve the style of their articles before submission [1]. However, these services are expensive, use translators who are not necessarily experts in the field, and are time-consuming, which often greatly delays the submission of articles [1].

Machine translators (MTs) are increasingly used in everyday life [2, 3]. Indeed, thanks to neural networks, the quality of translation has greatly improved in the last decades [46] and they do not require advanced computer skills. They are also used in medicine, for example to translate electronic medical records and to improve patient management in clinical practice, with mixed results [3, 715]. For example, Taira et al assessed the use of Google Translate for translating commonly used Emergency Department discharge instructions into seven languages [9]. While the overall meaning was retained in 82.5% of translations, accuracy rates varied across languages, and the study concluded that Google Translate should not be relied upon for patient instructions due to inconsistency in translation quality. In another study, Turner et al assessed the feasibility of using Google Translate followed by human post-editing to translate public health materials from English to Chinese [10]. The results showed that common machine translation errors and challenges in post-editing led to lower quality translations, suggesting the need for improvements in machine translation and post-editing processes before routine use in public health practice. However, a previous study of the same research team suggested that Google Translate and post-editing could yield translations of comparable quality in a more efficient and cost-effective manner for English to Spanish [11]. Blind ratings by two bilingual public health professionals indicated that when comparing human translation and machine translation followed by human post-editing, both types of translations were considered overall equivalent, with 33% preferring human translation, 33% preferring machine translation followed by human post-editing, and 33% finding both translations to be of equal quality. According to the authors, these divergent results between the two studies are linked to significant differences between English and Chinese, for example in syntactic structures. Khoong et al also found marked differences between Spanish and Chinese when using Google Translate for translations of emergency department discharge instructions [12]. Among the 100 sets of patient instructions containing 647 sentences, Google Translate accurately translated 92% of sentences into Spanish and 81% into Chinese. A minority of the inaccuracies in the translations had the potential for clinically significant harm.

Only a few studies evaluated the use of MTs in academic research, and they mainly focused on the extraction of relevant data from non-English articles [1618]. For example, Balk et al compared Google Translate’s ability to translate non-English language studies for systematic reviews in five languages and found variations in accuracy [16]. Spanish translations demonstrated the highest correct extraction rate (93% of items correctly extracted more than half the time), followed by German and Japanese (89%), French (85%), and Chinese (78%). According to the authors, caution is advised when using machine translation, as there is a trade-off between achieving comprehensive reviews and the potential for translation-related errors.

The objective of the current study was to assess the performance of three MTs, namely DeepL, Google Translate, and CUBBITT, in translating medical abstracts from French to English. We aimed to compare the accuracy of translations using nine metrics of Recall-Oriented Understudy for Gisting Evaluation (ROUGE), while also considering the stylistic quality through human evaluation. This study addressed the challenges faced by non-English speaking medical researchers and explored the practicality of using machine translation in this context. By testing our hypothesis that MTs may exhibit variations in translating medical research, we aimed to provide valuable insights for French-speaking researchers seeking to publish in English-language journals.

Methods

Selection of abstracts and machine translators (MTs)

We selected the two most prestigious general medical journals (according to the 2020 Journal Citation Reports impact factor) that translate all (Canadian Family Physician, impact factor = 3.3) or some (CMAJ, impact factor = 8.3) of the abstracts of published articles into French. We limited this preliminary study to general medical journals and did not include medical specialty or basic science journals that may use more technical language. We selected high-impact journals in a bilingual English/French country (Canada) to ensure that the French abstracts included in the study were of high quality.

We randomly extracted ten articles published in 2021 with abstracts available in French, five published in CMAJ (abstracts #1 to #5) and five in Canadian Family Physician (abstracts #6 to #10). We included ten articles in the study to obtain a variety of topics and study designs. Taken together, these ten abstracts contained 12,153 words in total.

Then, in spring 2022, we selected all MTs allowing the translation of at least 5,000 characters from French to English for free. Three MTs met these criteria (i.e., DeepL [https://www.deepl.com/translator], Google Translate [https://translate.google.com], and CUBBITT (Charles University Block-Backtranslation-Improved Transformer Translation) [https://lindat.mff.cuni.cz/services/translation]. At the time of the study, DeepL was free up to 5,000 characters, and 26 languages were available for translation; Google Translate was also free up to 5,000 characters, and over 100 languages were supported; CUBBITT had no character limit, but only six languages were available, including French and English.

Selection of metrics to evaluate the accuracy of the translation

We selected nine metrics of Recall-Oriented Understudy for Gisting Evaluation (ROUGE) [1921], namely ROUGE-1 recall/precision/F1-score, ROUGE-2 recall/precision/F1-score, and ROUGE-L recall/precision/F1-score.

ROUGE-N measures the number of identical n-grams between the text generated by a translator and a reference text considered as the gold standard. An n-gram is a grouping of words. For example, a unigram (1-gram) consists of a single word and a bigram (2-gram) consists of two consecutive words. The reference is a human-made optimal result. Thus, for ROUGE-1 and ROUGE-2, we measure the match rate of unigrams and bigrams, respectively, between the translated text and the reference. ROUGE-L measures the longest sequence of words that appear in the same order in both the translated text and the reference. The idea behind this metric is that a longer shared sequence indicates greater similarity between the two versions.

ROUGE-N and ROUGE-L are evaluated using three different metrics. The recall metric counts the number of identical n-grams, respectively the longest sequence of words that appear in both the translated text and the reference, divided by the total number of n-grams in the reference. It is used to verify that the translated text captures all the information contained in the reference. The precision metric is calculated in almost the same way, but, instead of dividing by the number of n-grams in the reference, it is divided by the number of n-grams in the translated text. It is used to check that the translator does not produce irrelevant words. Finally, the F1-score combines the recall and precision metrics to obtain an overall measure of translation accuracy.

Table 1 shows how to calculate the nine metrics using an example for the translated text and an example for the reference text. Implementing these metrics is easy in Python [20]. There are no recognized criteria defining above which scores of ROUGE a MT can be considered accurate. These measures are mainly used to compare MTs with each other, knowing that the higher the scores, the higher the accuracy of the translation. The main drawback of ROUGE is that it measures syntactic and not semantic matches. Thus, if two sequences have the same meaning, but use different words to express that meaning, the scores could be relatively low. For this reason, we also included a human evaluation of the translation performance, by analyzing the fluency score. This score is used to assess whether the text contains errors that native speakers would not have made or, more simply, whether the text is written in good English [22]. The best way to evaluate fluency is to use a multi-point fluency scale, with anchor text for each value [22]:

Table 1

Calculation of the different metrics of Recall-Oriented Understudy for Gisting Evaluation (ROUGE) using an example for the translated text and an example for the reference text.

MetricNumber of n-grams1 found in the translated text (“He eats drinks coffee muffins likes chocolate”)Number of n-grams found in the reference text (“He likes chocolate”)Number of n-grams (for ROUGE-1 and ROUGE-2), or longest common subsequence (for ROUGE-L), found in the translated text and the reference textRecall2Precision3F1-score4
ROUGE-15733 (i.e., “He”, “likes”, “chocolate”)3/3 = 13/7 = 0.432*(0.43*1)/(0.43+1) = 0.6
ROUGE-26621 (i.e., “likes chocolate”)1/2 = 0.51/6 = 0.172*(0.17*0.5)/(0.17+0.5) = 0.3
ROUGE-L7733 (i.e., “He likes chocolate”)3/3 = 13/7 = 0.432*(0.43*1)/(0.43+1) = 0.6

Open in a separate window

1 An n-gram is a grouping of words. For example, a unigram (1-gram) consists of a single word and a bigram (2-gram) consists of two consecutive words

2Recall = number of n-grams found in the translated text and the reference divided by number of n-grams in the reference

3Precision = number of n-grams found in the translated text and the reference divided by number of n-grams in the translated text

4F1-score = 2 * (precision * recall) / (precision + recall)

5ROUGE-1 measures the number of matching 1-grams between the translated text and the reference

6ROUGE-2 measures the number of matching 2-grams between the translated text and the reference

7ROUGE-L measures the longest sequence of words that appear in the same order in both the translated text and the reference. For ROUGE-L, we use the number of 1-grams for the denominator (number of 1-grams in the reference for recall and number of 1-grams in the translated text for precision)

How do you judge the fluency of the translation?

  • 5 → Flawless English

  • 4 → Good English

  • 3 → Non-Native English

  • 2 → Disfluent English

  • 1 → Incomprehensible

ROUGE and fluency were used in a large number of studies to evaluate texts. Koto et al. identified 106 studies using ROUGE and 45 studies examining fluency [23].

Data collection

We translated the ten abstracts from French to English using the three selected tools. The ten original abstracts in English and the versions obtained after translation by the three MTs are available elsewhere (https://doi.org/10.17605/OSF.IO/RCB36). Then, we evaluated the accuracy of the translation using the nine metrics of ROUGE, taking the original English abstract as reference text.

We also asked ten native English-speakers (five women and five men) to rate the fluency of the abstracts, including the original version, using the multi-point fluency scale. We added up the scores of the ten raters to get the overall score, ranging from 10 to 50. All study raters were acquaintances of the investigators, with a scientific background. In detail, there were five physicians and a fifth year medical student, two non-governmental organization (NGO) workers, a data scientist, and a manager. To avoid biasing the human evaluation, the raters were told that all versions were authored by translators in training, and the order of the versions was different for each abstract.

Statistical analyses

The nine metrics of ROUGE and the fluency score were first recorded separately for each version of each abstract. For ROUGE, the number of texts analyzed was 30 (i.e., the ten abstracts for each of the three MTs), whereas for the fluency score, this number was 40, since the original version of the ten abstracts were also analyzed. Using medians and interquartile ranges (IQRs), the results were then summarized for each of the MTs for ROUGE (n = 3), and for each of the MTs and the original version for the fluency score (n = 4). We summarized the results using medians and IQRs, because the data did not follow a normal distribution. We used Kruskal-Wallis tests to assess whether the differences in medians were statistically significant. The assumption of a similar distribution shape for all groups was met. If there were significant differences between groups, we used Dunn tests, with adjustments for multiple comparisons (Sidak), to identify the specific groups that differed from each other.

We also extracted the number of abstracts with the highest score for the nine metrics of ROUGE for DeepL, Google Translate, and CUBBITT respectively. We did the same for the fluency score, for DeepL, Google Translate, CUBBITT, and the reference text respectively.

Finally, for the human evaluation, we examined the inter-rater agreement between the ten raters for the reference text and the versions translated by DeepL, Google Translate, and CUBBITT. We used the ‘kappaetc’ command in Stata to calculate the quadratic weighted agreement coefficients (percent agreement and Gwet’s AC) [24, 25]. The statistical significance was set at a two-sided p-value of ≤0.05. All analyses were performed with STATA 15.1 (College Station, USA).

Ethical considerations

Since this study did not involve the collection of personal health-related data it did not require ethical review, according to current Swiss law.

Results

The raw data (i.e., the scores obtained for the nine metrics of ROUGE and the fluency score) are available in Open Science Framework (https://doi.org/10.17605/OSF.IO/RCB36).

Tables ​Tables22 (for ROUGE) and ​and33 (for the fluency score) summarize these data using medians and IQRs, as well as the number of abstracts with the highest score. Fig 1 presents the data in graphical form. ROUGE median scores were higher for Deepl than for CUBBITT, and higher for CUBBITT than for Google Translate, except for ROUGE-2 F1 and ROUGE-2 recall for which scores were higher for Google Translate than for CUBBITT. However, none of these differences was statistically significant (medians ranging from 0.5246 to 0.7392 for DeepL, from 0.4634 to 0.7200 for Google Translate, and from 0.4815 to 0.7316 for CUBBITT, all p-values > 0.10). For the overall fluency score (we added up the scores of the ten raters to get this score), CUBBITT tended to score higher than DeepL, Google Translate, and the original English text (median = 43 for CUBBITT, vs. 39 for DeepL, 38 for Google Translate, and 40 for the original English text, p-value = 0.003). The difference in median score was borderline significant between CUBBITT and the reference text (p-value = 0.05), whereas it was statistically significant between CUBBITT and DeepL (p-value = 0.03), and between CUBBITT and Google Translate (p-value = 0.001). All ten abstracts received an individual score ranging from 3 to 5 (no score was below 3), even Google Translate, which achieved the lowest overall median score.

Open in a separate window

Fig 1

Recall-Oriented Understudy for Gisting Evaluation (ROUGE) and fluency median scores.

Data are presented for the three machine translators (MTs) separately for ROUGE, and for the three MTs and the reference text for the fluency score. Median scores for ROUGE are presented as percentages.

Table 2

Median score (IQR) and number of abstracts with the highest score for the nine Recall-Oriented Understudy for Gisting Evaluation (ROUGE) metrics used to evaluate translations of ten abstracts of medical scientific articles by three machine translators (DeepL, Google Translate, and CUBBITT).

Machine translator (MT)ROUGE-1 F1-scoreROUGE-1 precisionROUGE-1 recallROUGE-2 F1-scoreROUGE-2 precisionROUGE-2 recallROUGE-L F1-scoreROUGE-L precisionROUGE-L recall
DeepL
Median score (IQR) for the ten abstracts0,7352 (0.1298)0,7312 (0,1077)0,7392 (0,1524)0,5259 (0,1700)0,5273 (0.1743)0,5246 (0.1677)0,7241 (0.1370)0,7202 (0.1122)0,7281 (0.1625)
Number (%) of abstracts with the highest score6 (60.0)7 (70.0)7 (70.0)8 (80.0)9 (90.0)7 (70.0)6 (60.0)8 (80.0)6 (60.0)
Google Translate
Median score (IQR) for the ten abstracts0,6978 (0.0740)0,6817 (0.0549)0,7200 (0.0881)0,4845 (0.0804)0,4634 (0.0921)0,4947 (0.0852)0,6856 (0.0933)0,6738 (0.0779)0,7073 (0.1047)
Number (%) of abstracts with the highest score1 (10.0)02 (20.0)001 (10.0)2 (20.0)02 (20.0)
CUBBITT
Median score (IQR) for the ten abstracts0,7264 (0.1273)0,7214 (0.0990)0,7316 (0.1349)0,4828 (0.1456)0,4815 (0.1425)0,4846 (0.1488)0,7084 (0.1433)0,6995 (0.1103)0,7175 (0.1513)
Number (%) of abstracts with the highest score3 (30.0)3 (30.0)1 (10.0)2 (20.0)1 (10.0)2 (20.0)2 (20.0)2 (20.0)2 (20.0)
p-value 10.590.180.700.430.270.680.620.300.76

Open in a separate window

1 Kruskal-Wallis tests to assess whether differences in median scores between the three MTs were statistically significant

Table 3

Median fluency score (IQR) and number of abstracts with the highest score.

This score was used to assess the style of ten original English abstracts (= reference text) and the versions translated by DeepL, Google Translate, and CUBBITT.

Reference text or machine translator (MT)Fluency score1,2
Median (IQR) or N (%)
Reference text
Median score (IQR) for the ten abstracts40 (4)
Number (%) of abstracts with the highest score0
DeepL
Median score (IQR) for the ten abstracts39 (3)
Number (%) of abstracts with the highest score2 (20)
Google Translate
Median score (IQR) for the ten abstracts38 (1)
Number (%) of abstracts with the highest score0
CUBBITT
Median score (IQR) for the ten abstracts43 (4)
Number (%) of abstracts with the highest score8 (80)

Open in a separate window

1 The evaluation was done independently by ten native English-speaking raters. They were asked to answer to the question “How do you judge the fluency of the translation?” Possible answers were 5 (= flawless English), 4 (= good English), 3 (= non-native English), 2 (= disfluent English), 1 (= incomprehensible). We added up the scores of the ten raters to get the overall score, ranging from 10 to 50.

2 Kruskal-Wallis or Dunn tests to assess whether differences in median scores between the groups were statistically significant. P-value for the difference between the four groups = 0.003, between CUBBITT and the reference text = 0.05, between CUBBITT and DeepL = 0.03, between CUBBITT and Google Translate = 0.001, between the reference text and DeepL = 0.95, between the reference text and Google Translate = 0.45, between DeepL and Google Translate = 0.62.

ROUGE scores were highest for six to nine abstracts for DeepL (depending on the score considered), one to three abstracts for CUBBITT, and zero to two abstracts for Google Translate. Fluency scores were highest for eight abstracts for CUBBITT, two abstracts for DeepL, and zero abstracts for Google Translate and the reference text.

Finally, the inter-rater agreement between the ten raters was high (Table 4). The raters agreed on more than 97% of the abstracts (p-values < 0.001), and the chance-corrected Gwet’s agreement coefficients were high (p-values < 0.001).

Table 4

Inter-rater agreement between the ten native English-speaking raters who evaluated the style of ten original English abstracts (= reference text) and the versions translated by DeepL, Google Translate, and CUBBITT.

Reference text or machine translator (MT)Percent agreement (95%CI)p-valueGwet’s AC (95%CI)p-value
Reference text0.9789 (0.9691–0.9886)<0.0010.9667 (0.9437–0.9897)<0.001
DeepL0.9792 (0.9696–0.9887)<0.0010.9684 (0.9476–0.9892)<0.001
Google Translate0.9786 (0.9699–0.9873)<0.0010.9665 (0.9466–0.9865)<0.001
CUBBITT0.9739 (0.9679–0.9799)<0.0010.9530 (0.9364–0.9696)<0.001

Open in a separate window

The ten raters were asked to answer to the question “How do you judge the fluency of the translation?” Possible answers were 5 (= flawless English), 4 (= good English), 3 (= non-native English), 2 (= disfluent English), 1 (= incomprehensible). We calculated the quadratic weighted agreement coefficients (percent agreement and Gwet’s AC)

Discussion

Main findings

We compared the performance of three MTs (i.e., DeepL, Google Translate, and CUBBITT) for translating medical research abstracts from French to English. In this preliminary study, we evaluated five abstracts published in CMAJ and five abstracts published in Canadian Family Physician. We found that the three MTs performed similarly when tested with ROUGE, but CUBBITT was slightly better than the other two using human evaluation. We also found that for human evaluation CUBBITT tended to perform better than the original English text.

Comparison with existing literature

MTs are increasingly used in medicine, particularly to translate electronic medical records and improve patient care [3, 715]. Current evidence also suggests that they are relatively reliable for extracting data from non-English articles in systematic reviews [16, 17]. However, there is little data available on their effectiveness in academic research. Using a subjective evaluation method, Takakusagi et al. investigated the accuracy of DeepL in translating an entire medical article from Japanese into English [26]. The authors compared the original Japanese article with the English version, which was back translated into Japanese by medical translators. They found that the overall accuracy was high, with an average match rate of 94%. However, the accuracy varied between sections of the article, with the ’Results’ section showing the highest accuracy (100%) and the ’Materials and Methods’ section showing the lowest accuracy (89%). The authors limited their analysis to the accuracy of the meanings and did not assess the stylistic quality of the translation.

In our study, we found that DeepL, Google Translate, and CUBBITT are three effective tools for accurately and fluently translating abstracts of medical articles from French into English. Surprisingly, CUBBITT did better in terms of fluency as the original English abstracts published in two high-impact English-language journals. These results are however in line with a recent study conducted by the developers of CUBBITT, which showed that the quality of translations done with this tool approached that of professional translators in terms of fluency [27].

Unlike the biomedical sciences, a large amount of data on machine translation is available in the field of educational linguistics, second language studies, and foreign language education. Two review articles have recently been published [28, 29]. These papers summarized the key concepts, insights, and findings, categorizing them into questions like how learners use MTs, what instructors and learners think about MTs, and how MTs affect language learning. Students have diverse opinions concerning the appropriateness, reliability, and ethical considerations of machine translation tools [3033]. Learners generally hold favorable views of machine translation, believing it has the potential to assist their learning and enhance the quality of their second language writing. However, these positive perceptions are counterbalanced by concerns about machine translation accuracy, an understanding of its limitations, and conflicting interpretations of what constitutes ethical behavior. The literature exploring the potential advantages of machine translation in language learning did not produce definitive findings. However, it suggests two potential trends: MTs might serve as a valuable resource for improving learners’ metalinguistic understanding [3437], and they can aid students in achieving better results in translation and second language writing tasks [38, 39]. Some of these studies focused on the use of Google Translate [4042], yet neither review included studies that made direct comparisons between MTs.

To our knowledge, few comparative studies are available in the literature. Hidalgo-Ternero assessed the performance of Google Translate and DeepL in translating Spanish idiomatic expressions into English, including both common and less frequent variants, with a focus on whether these idioms were presented in continuous or discontinuous forms [43]. The study found that Google Translate and DeepL performed well in accurately translating high-frequency idiomatic expressions, achieving an average accuracy rate of 86% and 89%, respectively. However, they struggled to detect and translate lower-frequency phraseological variants of these idioms, indicating limitations in handling less common idiomatic expressions. Focusing on human post-editing efforts, another study compared the performance of three MTs for translating Cochrane plain language health information from English to Russian [13]. The authors found that Google Translate performed best, slightly better than DeepL, while Microsoft Translator performed less well.

Our study was not designed to estimate the amount of time needed by researchers for post-editing (i.e., the time needed to make corrections to the text after it has been translated into English by the MT). Given the results obtained for the fluency score, post-editing should nevertheless be performed fairly quickly. Indeed, even without post-editing, the stylistic quality of the translation was considered by the evaluators to be better (for CUBBITT) or almost as good (for DeepL and Google Translate) as the original text.

This preliminary study included only abstracts, which are generally written to be more accessible and more quickly “digested” than full articles. We did not evaluate the performance of MTs with full articles. Translation apps often lack specialized medical terminology, which can make them useless for translating highly specialized medical articles. Further studies evaluating the performance of MTs for full articles and for various disciplines are therefore needed. However, we believe that non-English-speaking researchers who do not wish to rely on the services of professional translators (e.g. because of their cost) could have an interest in using DeepL, Google Translate, or CUBBITT for some of their work that is not highly specialized. Indeed, the time spent in post-editing after using these MTs probably be far outweighed by the time they would have to spend translating scientific articles themselves or the time spent writing the articles directly in English.

Strengths and limitations

Our study has several strengths. We incorporated a dual assessment approach, combining quantitative ROUGE metrics and qualitative fluency evaluations by native English speakers. This ensures a comprehensive evaluation of machine translation tools, providing a nuanced understanding of both syntactic and semantic aspects of the translations. In addition, focusing on medical texts, our method aligns with practical scenarios faced by non-English-speaking researchers. By evaluating tools in a domain-specific context, our approach offers insights directly applicable to researchers in the medical field, enhancing the relevance of the study. Finally, the inclusion of raters with varied scientific backgrounds enhances the robustness of the fluency assessment. This diversity ensures a broad perspective on the quality of translations, considering the expectations and language nuances across different professional domains.

However, our study also has some weaknesses. First, we included only French abstracts published in two general medical journals. It is not certain that the results would have been similar for full articles, other languages, and/or other journals. The selection of two bilingual journals introduces a potential limitation as the study’s outcomes rely on the quality of translated abstracts from English to French published in these journals. While the stylistic quality of these versions was generally deemed good or excellent by the raters, it is essential to acknowledge the influence of the initial French abstracts on the translation process. Second, although ROUGE is a validated instrument that is often used to evaluate the performance of MTs, it does not measure semantic matches. If two sequences have the same meaning, but use different words to express that meaning, the score assigned could be relatively low. Third, only ten abstracts were included in the study and only ten raters were recruited for the human evaluation. We included only ten abstracts, because it was important for the evaluators to carefully assess the four versions of the abstracts (the original version and the versions from DeepL, Google Translate, and CUBBITT), and this was a time-consuming task. Future studies may consider including a larger sample to obtain more robust results. Finally, we selected abstracts from the year 2021 to ensure that the texts were current and reflected the latest developments in medicine. Future studies may encompass a broader time frame to examine variations over the years.

Conclusion

Our study provides a thorough examination of the performance of MTs—DeepL, Google Translate, and CUBBITT—in the specific context of translating medical research abstracts from French to English. This focused evaluation contributes to a nuanced understanding of the applicability of these tools in the medical domain. We not only assessed the accuracy of translations using established metrics but also delved into the fluency of the translated text. Our study aims to highlight the practical utility of MTs for non-English-speaking researchers in medicine.

We found that the three MTs performed similarly when tested with ROUGE, but CUBBITT was slightly better than the other two using human evaluation. We also found that in terms of stylistic quality CUBBITT tended to perform better than the original English text.

Although the study was limited to the analysis of abstracts published in general medical journals and did not evaluate the time required for post-editing, we believe that French-speaking researchers could benefit from using DeepL, Google Translate, or CUBBITT to translate articles written in French into English. Further studies would be needed to evaluate the performance of MTs with full articles and languages other than French.

Funding Statement

The authors received no specific funding for this work.

Data Availability

Data availability statement: The data associated with this article are available in Open Science Framework (https://doi.org/10.17605/OSF.IO/RCB36).

References

1. Ramírez-Castañeda V.Disadvantages in preparing and publishing scientific papers caused by the dominance of the English language in science: The case of Colombian researchers in biological sciences.PLOS ONE. 2020. Sep16;15(9):e0238372. doi: 10.1371/journal.pone.0238372 [PMC free article] [PubMed] [CrossRef] [Google Scholar]

2. Vieira LN, O’Sullivan C, Zhang X, O’Hagan M. Machine translation in society: insights from UK users.Lang Resour Eval.2023. Jun1;57(2):893–914. [Google Scholar]

3. Dew KN, Turner AM, Choi YK, Bosold A, Kirchhoff K. Development of machine translation technology for assisting health communication: A systematic review.J Biomed Inform. 2018. Sep;85:56–67. doi: 10.1016/j.jbi.2018.07.018 [PubMed] [CrossRef] [Google Scholar]

4. Hirschberg J, Manning CD. Advances in natural language processing. Science. 2015. Jul17;349(6245):261–6. doi: 10.1126/science.aaa8685 [PubMed] [CrossRef] [Google Scholar]

5. Song R.Analysis on the Recent Trends in Machine Translation. Highlights Sci Eng Technol.2022. Nov10;16:40–7. [Google Scholar]

6. Mondal SK, Zhang H, Kabir HMD, Ni K, Dai HN. Machine translation and its evaluation: a study.Artif Intell Rev.2023. Sep1;56(9):10137–226. [Google Scholar]

7. Soto X, Perez-de-Viñaspre O, Labaka G, Oronoz M. Neural machine translation of clinical texts between long distance languages.J Am Med Inform Assoc JAMIA.2019. Dec1;26(12):1478–87. doi: 10.1093/jamia/ocz110 [PMC free article] [PubMed] [CrossRef] [Google Scholar]

8. Randhawa G, Ferreyra M, Ahmed R, Ezzat O, Pottie K. Using machine translation in clinical practice.Can Fam Physician.2013. Apr;59(4):382–3. [PMC free article] [PubMed] [Google Scholar]

9. Taira BR, Kreger V, Orue A, Diamond LC. A Pragmatic Assessment of Google Translate for Emergency Department Instructions.J Gen Intern Med. 2021. Nov;36(11):3361–5. doi: 10.1007/s11606-021-06666-z [PMC free article] [PubMed] [CrossRef] [Google Scholar]

10. Turner AM, Dew KN, Desai L, Martin N, Kirchhoff K. Machine Translation of Public Health Materials From English to Chinese: A Feasibility Study.JMIR Public Health Surveill.2015;1(2):e17. doi: 10.2196/publichealth.4779 [PMC free article] [PubMed] [CrossRef] [Google Scholar]

11. Turner AM, Bergman M, Brownstein M, Cole K, Kirchhoff K. A comparison of human and machine translation of health promotion materials for public health practice: time, costs, and quality.J Public Health Manag Pract JPHMP. 2014;20(5):523–9. doi: 10.1097/PHH.0b013e3182a95c87 [PMC free article] [PubMed] [CrossRef] [Google Scholar]

12. Khoong EC, Steinbrook E, Brown C, Fernandez A. Assessing the Use of Google Translate for Spanish and Chinese Translations of Emergency Department Discharge Instructions.JAMA Intern Med. 2019. Apr;179(4):580–2. doi: 10.1001/jamainternmed.2018.7653 [PMC free article] [PubMed] [CrossRef] [Google Scholar]

13. Ziganshina LE, Yudina EV, Gabdrakhmanov AI, Ried J. Assessing Human Post-Editing Efforts to Compare the Performance of Three Machine Translation Engines for English to Russian Translation of Cochrane Plain Language Health Information: Results of a Randomised Comparison.Informatics. 2021. Mar;8(1):9. [Google Scholar]

14. Dahal SB, Aoun M. Exploring the Role of Machine Translation in Improving Health Information Access for Linguistically Diverse Populations.Adv Intell Inf Syst.2023. Apr3;8(2):1–13. [Google Scholar]

15. Herrera-Espejel PS, Rach S. The Use of Machine Translation for Outreach and Health Communication in Epidemiology and Public Health: Scoping Review.JMIR Public Health Surveill.2023. Nov20;9:e50814. doi: 10.2196/50814 [PMC free article] [PubMed] [CrossRef] [Google Scholar]

16. Balk EM, Chung M, Chen ML, Chang LKW, Trikalinos TA. Data extraction from machine-translated versus original language randomized trial reports: a comparative study.Syst Rev.2013. Nov7;2:97. doi: 10.1186/2046-4053-2-97 [PMC free article] [PubMed] [CrossRef] [Google Scholar]

17. Jackson JL, Kuriyama A, Anton A, Choi A, Fournier JP, Geier AK, et al. The Accuracy of Google Translate for Abstracting Data From Non-English-Language Trials for Systematic Reviews. Ann Intern Med. 2019. Nov5;171(9):677–9. doi: 10.7326/M19-0891 [PubMed] [CrossRef] [Google Scholar]

18. Zulfiqar S, Wahab MF, Sarwar MI, Lieberwirth I. Is Machine Translation a Reliable Tool for Reading German Scientific Databases and Research Articles?J Chem Inf Model.2018. Nov26;58(11):2214–23. doi: 10.1021/acs.jcim.8b00534 [PubMed] [CrossRef] [Google Scholar]

19. Lin CY, Och FJ. Automatic evaluation of machine translation quality using longest common subsequence and skip-bigram statistics. In: Proceedings of the 42nd Annual Meeting on Association for Computational Linguistics—ACL ‘04 [Internet].Barcelona, Spain: Association for Computational Linguistics; 2004. [cited 2023 Nov 5]. p. 605–es. Available from: http://portal.acm.org/citation.cfm?doid=1218955.1219032 [Google Scholar]

20. Briggs J.The Ultimate Performance Metric in NLP [Internet]. Medium. 2021. [cited 2023 Nov 5]. Available from: https://towardsdatascience.com/the-ultimate-performance-metric-in-nlp-111df6c64460 [Google Scholar]

21. Lin CY. ROUGE: A Package for Automatic Evaluation of Summaries.In: Text Summarization Branches Out [Internet]. Barcelona, Spain: Association for Computational Linguistics; 2004. p. 74–81. Available from: https://aclanthology.org/W04-1013 [Google Scholar]

22. Olive J, Christianson C, McCary J. Handbook of natural language processing and machine translation: DARPA global autonomous language exploitation. New York: Springer; 2011. [Google Scholar]

23. Koto F, Baldwin T, Lau JH. FFCI: A Framework for Interpretable Automatic Evaluation of Summarization. J Artif Intell Res [Internet].2022. Apr29 [cited 2023 Nov 5];73. Available from: https://www.jair.org/index.php/jair/article/view/13167 [Google Scholar]

24. Klein D.Implementing a General Framework for Assessing Interrater Agreement in Stata. Stata J.2018. Dec1;18(4):871–901. [Google Scholar]

25. Gwet KL. Handbook of inter-rater reliability: the definitive guide to measuring the extent of agreement among raters. Fourth edition. Gaithersburg, Md: Advances Analytics, LLC; 2014. 410 p. [Google Scholar]

26. Takakusagi Y, Oike T, Shirai K, Sato H, Kano K, Shima S, et al. Validation of the Reliability of Machine Translation for a Medical Article From Japanese to English Using DeepL Translator.Cureus. 2021. Sep;13(9):e17778. [Google Scholar]

27. Popel M, Tomkova M, Tomek J, Kaiser Ł, Uszkoreit J, Bojar O, et al. Transforming machine translation: a deep learning system reaches news translation quality comparable to human professionals.Nat Commun.2020. Sep1;11(1):4381. doi: 10.1038/s41467-020-18073-9 [PMC free article] [PubMed] [CrossRef] [Google Scholar]

28. Jolley JR, Maimone L. Thirty Years of Machine Translation in Language Teaching and Learning: A Review of the Literature. L2 J Electron Refereed J Foreign Second Lang Educ [Internet].2022. [cited 2023 Nov 5];14(1). Available from: https://escholarship.org/uc/item/2df5d55c [Google Scholar]

29. Urlaub P, Dessein E. Machine translation and foreign language education.Front Artif Intell. 2022;5:936111. doi: 10.3389/frai.2022.936111 [PMC free article] [PubMed] [CrossRef] [Google Scholar]

30. Jin L, Deifell E. Foreign Language Learners’ Use and Perception of Online Dictionaries: A Survey Study. 2013;9(4). [Google Scholar]

31. Maimone L, Jolley J. Free Online Machine Translation: Use and Perceptions by Spanish Students and Instructors. [cited 2023 Nov 5]; Available from: https://www.academia.edu/11355802/Free_Online_Machine_Translation_Use_and_Perceptions_by_Spanish_Students_and_Instructors [Google Scholar]

32. Larson-Guenette J. “It’s just reflex now”: German Language Learners’ Use of Online Resources. Unterrichtspraxis Teach Ger.2013;46(1):62–74. [Google Scholar]

33. Niño A.Exploring the use of online machine translation for independent language learning. Res Learn Technol [Internet].2020. Sep4 [cited 2023 Nov 5];28. Available from: https://journal.alt.ac.uk/index.php/rlt/article/view/2402 [Google Scholar]

34. Benda J.Google Translate in the EFL Classroom: Taboo or Teaching Tool?Writ Pedagogy.2013;5(2):317–32. [Google Scholar]

35. Enkin E, Mejías-Bikandi E. Using online translators in the second language classroom: Ideas for advanced-level Spanish. Lat Am J Content Lang Integr Learn [Internet].2016. Jun11 [cited 2023 Nov 5];9(1). Available from: https://laclil.unisabana.edu.co/index.php/LACLIL/article/view/6007 [Google Scholar]

36. Niño A.Evaluating the use of machine translation post-editing in the foreign language class.Comput Assist Lang Learn.2008. Feb1;21(1):29–49. [Google Scholar]

37. Correa M.Leaving the “peer” out of peer-editing: Online translators as a pedagogical tool in the Spanish as a second language classroom.Lat Am J Content Lang Integr Learn.2014. Apr30;7(1):1–20. [Google Scholar]

38. Garcia I, Pena MI. Machine Translation-Assisted Language Learning: Writing for Beginners.Comput Assist Lang Learn. 2011;24(5):471–87. [Google Scholar]

39. Fredholm K.Effects of online translation on morphosyntactic and lexical-pragmatic accuracy in essay writing in Spanish as a foreign language.In: CALL Design: Principles and Practice—Proceedings of the 2014 EUROCALL Conference, Groningen, The Netherlands [Internet].Research-publishing.net; 2014. [cited 2023 Nov 5]. p. 96–101. Available from: https://research-publishing.net/manuscript?10.14705/rpnet.2014.000201 [Google Scholar]

40. Kol S, Schcolnik M, Spector-Cohen E. Google Translate in Academic Writing Courses?EuroCALL Rev.2018. Sep30;26(2):50–7. [Google Scholar]

41. Fredholm K.Efectos del traductor de Google sobre la diversidad léxica: el desarrollo de vocabulario entre estudiantes de español como lengua extranjera: Effects of Google translate on lexical diversity: vocabulary development among learners of Spanish as a foreign language.Rev Nebrija Lingüíst Apl Enseñ Leng. 2019. Apr29;13(26):98–117. [Google Scholar]

42. Lee SM. The impact of using machine translation on EFL students’ writing. Comput Assist Lang Learn.2020. Mar3;33(3):157–75. [Google Scholar]

43. Hidalgo Ternero C.Google Translate vs. DeepL: analysing neural machine translation performance under the challenge of phraseological variation.MonTI Monogr Trad E Interpret. 2021. Jan13;154–77. [Google Scholar]

Articles from PLOS ONE are provided here courtesy of PLOS

Performance of machine translators in translating French medical research abstracts to English: A comparative study of DeepL, Google Translate, and CUBBITT (2024)

FAQs

How does DeepL compare to Google Translate? ›

Is DeepL better than Google Translate? It depends on what you need. For the languages that it supports, DeepL is regarded to be a bit more accurate than Google Translate. However, DeepL Translate has a limited language selection as compared to Google Translate and also costs more money.

Is ChatGPT better than DeepL? ›

Differences between ChatGPT and DeepL

Unlike ChatGPT, DeepL was explicitly designed and trained for translation. For this reason, its translations are currently better than those of the chatbot, particularly when it comes to more complex texts.

Is DeepL good for translating? ›

As long as DeepL supports your language, the main differences are this: DeepL is generally rated as having higher-quality translations, along with options to choose between formal and informal tones. So if you want the most accurate and natural-sounding translations, DeepL is usually the best option.

What is the most accurate translator for French? ›

Google Translate

Its French-to-English translations are an impressive 97% accurate, so you can feel (mostly) confident when using this app! Google Translate is most accurate when translating short, simple sentences, so avoid long text or overly complicated language when using this app.

What translator is 100% accurate? ›

DeepL Translate: The world's most accurate translator.

Is there a machine translation better than Google? ›

Also known as DeepL Translator or DeepL.com, DeepL prides itself on delivering “the world's best machine translation.” Its proprietary neural networks are trained with the Linguee database to identify even the smallest nuances in text and generate the highest-quality translation possible.

Is DeepL no longer free? ›

Both private and business use of the DeepL Translator (free) and DeepL Write (free) services (together “Free services”) are free of charge. However, forwarding our Free Services to third parties for payment is strictly prohibited.

What is the best most accurate translator? ›

4 Best Translation Tools
  • Google Translate. Google Translate is a free internet-based machine translation tool that translates texts, documents, and websites into different languages. ...
  • Bing Microsoft Translator. ...
  • MemoQ Translator PRO. ...
  • Memsource.

What is the most advanced machine translation? ›

DeepL Translator

Thanks to its use of artificial intelligence, DeepL Translator can deliver more precise and nuanced translations compared to its peers. It may truly live up to its own claim of being “the world's most accurate translator.”

What are the drawbacks of DeepL? ›

Translation problems in very specific cases or for less common languages. The biggest downside to DeepL pro is the cost. It is much more expensive than many other translation services, and the monthly fee structure can be prohibitive for many users.

Is DeepL 100% accurate? ›

In blind tests, professional translators select the most accurate translation without knowing which company produced it. DeepL outperforms competition by a factor of 3:1. When talking about our best-in-class translation quality, it's important for us to back up our words with data.

Does DeepL count as AI? ›

DeepL Write: AI-powered writing companion.

Is reverso or DeepL better? ›

DeepL Pro has 135 reviews and a rating of 4.76 / 5 stars vs Reverso Context which has 33 reviews and a rating of 4.45 / 5 stars.

What is the best website to translate from English to French? ›

A top-tier English to French translator from QuillBot. Translate entire documents from English to French or from French to English in a matter of seconds with QuillBot's top-tier translation tool.

What is the best AI translator for French? ›

1. DeepL. DeepL Translator has gained acclaim for delivering some of the most accurate translations, especially in European languages. It supports over 30 languages, including English, German, French, Spanish, Chinese, Polish, Turkish, and Indonesian.

Which is better than DeepL translation? ›

Other important factors to consider when researching alternatives to DeepL include documents. The best overall DeepL alternative is Google Translate. Other similar apps like DeepL are Azure Translator Text API, Bing Translator, Microsoft Translator, and Language Weaver.

How accurate is DeepL percentage? ›

Reference text or machine translator (MT)Percent agreement (95%CI)Gwet's AC (95%CI)
Reference text0.9789 (0.9691–0.9886)0.9667 (0.9437–0.9897)
DeepL0.9792 (0.9696–0.9887)0.9684 (0.9476–0.9892)
Google Translate0.9786 (0.9699–0.9873)0.9665 (0.9466–0.9865)
CUBBITT0.9739 (0.9679–0.9799)0.9530 (0.9364–0.9696)
Feb 1, 2024

Will DeepL replace translators? ›

- DeepL can't fully replace the skill and creativity of human translation. It may miss cultural nuances, context, and tone, or produce some wonky translations.

Top Articles
Latest Posts
Article information

Author: Mrs. Angelic Larkin

Last Updated:

Views: 6052

Rating: 4.7 / 5 (67 voted)

Reviews: 82% of readers found this page helpful

Author information

Name: Mrs. Angelic Larkin

Birthday: 1992-06-28

Address: Apt. 413 8275 Mueller Overpass, South Magnolia, IA 99527-6023

Phone: +6824704719725

Job: District Real-Estate Facilitator

Hobby: Letterboxing, Vacation, Poi, Homebrewing, Mountain biking, Slacklining, Cabaret

Introduction: My name is Mrs. Angelic Larkin, I am a cute, charming, funny, determined, inexpensive, joyous, cheerful person who loves writing and wants to share my knowledge and understanding with you.