Cautiously, cautiously...
Aug. 3rd, 2002 01:42 pmLJ friend
vuzh comments:
Machine translation attempts to translate from source to target using a dictionary and some complex collection of rules that try to figure out what the original means. As you point out, the result is generally crappy.
Translation memory programs, on the other hand, rely on a database of previous (presumably human) translations.
For example, if - in a previous translation - the sentence "La plume de ma tante est ici" is rendered as "My aunt's pen is here," then upon encountering the same sentence, the program fetches the translation and inserts it into the target.
If the program encounters the sentence "La plume de mon oncle est ici," it will typically insert the previous translation ("My aunt's pen..."),, highlight "mon oncle," and indicate that the translation is 80% (or whatever number it calculates) identical to a previously encountered sentence.
The translator then reviews the sentence, translates "mon oncle," and makes any other required changes, and goes on to the next sentence. It's all pretty cut-and-dried and the program makes no pretense of knowing more than you about what needs doing. You can think of TM as a very sophisticated, but dumb pattern-matcher.
TM programs are increasingly popular among translators, since they (a) help assure consistency and (b) help make sure no sentences are omitted from the translation. They are also a tremendous boon when it comes to translating text with a lot of boilerplate, because once you translate such text once, it is propagated throughout the document (and subsequent documents, too), and any small, subtle differences are highlighted, so your attention can focus on the differences instead of rereading the rest of the text.
The down side of TM programs is that agencies are using them increasingly, too. I used to participate on a translator's mailing list - LANTRA-L - until the noise-to-signal ratio shot through the roof, but recall a number of discussions about how agencies would send freelancers databases of pre-translated text for TM programs such as Trados or Déjà Vu, and basically impose the same kinds of conditions as I'm seeing with Client U (full pay for new text; partial pay for stuff that's been translated before).
Of course, everything you send back in as new text is subsequently added to the database, so that next time, there's a broader range of text that needs no translation. Technically, one might assume that eventually the translator's job will be eliminated, but that's not quite accurate.
What is accurate is that the quantity of text that needs to be verified goes up. Moreover, unless I miss my guess, what you do with such text is not so much "editing" as it is "tweaking" (but I suspect I'll be better qualified to address this point after I do this project).
Talking about doing this project... I should probably get back to work. :^)
Cheers...
![[livejournal.com profile]](https://www.dreamwidth.org/img/external/lj-userinfo.gif)
It seems they are using some kind of Translation Memory program to pre-translate electronic files. The deal is that I get full freight for material that's not been pre-translated, but only 1/3 of my rate for "editing" the rest.Well, it's not as bad as all that.
i've done a lot of this type of thing with spanish translations.
make them send the original text as well --
those translation programs screw the text up so much, it's often difficult to impossible to figure out what the original said.
next time, you'll probably want to insist on a higher rate for "editing".
maybe someday automatic translation programs will work well, but right now, they're crap.
Machine translation attempts to translate from source to target using a dictionary and some complex collection of rules that try to figure out what the original means. As you point out, the result is generally crappy.
Translation memory programs, on the other hand, rely on a database of previous (presumably human) translations.
For example, if - in a previous translation - the sentence "La plume de ma tante est ici" is rendered as "My aunt's pen is here," then upon encountering the same sentence, the program fetches the translation and inserts it into the target.
If the program encounters the sentence "La plume de mon oncle est ici," it will typically insert the previous translation ("My aunt's pen..."),, highlight "mon oncle," and indicate that the translation is 80% (or whatever number it calculates) identical to a previously encountered sentence.
The translator then reviews the sentence, translates "mon oncle," and makes any other required changes, and goes on to the next sentence. It's all pretty cut-and-dried and the program makes no pretense of knowing more than you about what needs doing. You can think of TM as a very sophisticated, but dumb pattern-matcher.
TM programs are increasingly popular among translators, since they (a) help assure consistency and (b) help make sure no sentences are omitted from the translation. They are also a tremendous boon when it comes to translating text with a lot of boilerplate, because once you translate such text once, it is propagated throughout the document (and subsequent documents, too), and any small, subtle differences are highlighted, so your attention can focus on the differences instead of rereading the rest of the text.
The down side of TM programs is that agencies are using them increasingly, too. I used to participate on a translator's mailing list - LANTRA-L - until the noise-to-signal ratio shot through the roof, but recall a number of discussions about how agencies would send freelancers databases of pre-translated text for TM programs such as Trados or Déjà Vu, and basically impose the same kinds of conditions as I'm seeing with Client U (full pay for new text; partial pay for stuff that's been translated before).
Of course, everything you send back in as new text is subsequently added to the database, so that next time, there's a broader range of text that needs no translation. Technically, one might assume that eventually the translator's job will be eliminated, but that's not quite accurate.
What is accurate is that the quantity of text that needs to be verified goes up. Moreover, unless I miss my guess, what you do with such text is not so much "editing" as it is "tweaking" (but I suspect I'll be better qualified to address this point after I do this project).
Talking about doing this project... I should probably get back to work. :^)
Cheers...