Skip to main content

Artificial Intelligence and Translation Technology

The brief evolution of Machine Translation

The first machine translation experiment started in 1954 at IBM. In 1966 machine translation was declared as "expensive, inaccurate, and unpromising". This attitude slowed research for almost a decade. The first attempt at machine translation was a rule-based model. The RBMT had input sentences in a source language, which the system then generated as output sentences in a target language on the basis of morphological, syntactic, and semantic analysis of both the source and the target languages involved in a given translation task. But rule-based translation is extremely difficult for certain languages such as Japanese.

Example-based machine translation first was suggested in 1984 by Makoto Nagao from Kyoto University. The foundation of the EBMT is the idea of translation by analogy. Example-based translations or phrasal translations are translated by analogy to previous translations and reject the idea that human translators execute translations by doing a deep linguistic analysis. EBMT finally showed some light to researchers; they realized that instead of feeding the machine with endless linguistic and grammar rules and exceptions, they could use existing translations as examples.

Just five years later Statistical Machine Translation appeared in the early '90s. This machine translation did not know anything about rules, rather it analyzed two similar texts and tried to find the pattern. The machine did not rely on any dictionaries or grammar books but was simply statistical. The more text was used, the more accurate the translation became. SMT's biggest achievement was phrase-based translation. By 2006, every big tech company - from Google to Yandex and Bing - started to use this method. If you recall the instances when your Google translation results were either very on-point or were total nonsense, that was phrase-based machine translation.

For a very short time, syntax-based translation was thought to be the "next thing" but combining the syntax-based approach with the rule-based approach was not successful.

In 2014 the first news about Neural Machine Translation came out. In 2016 Google made a game-changing announcement. Neural Machine Translation works based on encoding one thing on one side and decoding it at the other end. In two years, neural networks surpassed everything that had appeared in the past 20 years of translation. Neural translation contains 50% fewer word order mistakes, 17% fewer lexical mistakes, and 19% fewer grammar mistakes.

Statistical machine translation methods always worked using English as the key source. Thus, if you translated from Russian to German, the machine first translated the text to English and then from English to German. The loss was double.

Neural translation on the other hand doesn’t need that — only a decoder is required for it to work. That was the first time that direct translation between languages with no сommon dictionary became possible. In 2016, Google turned on neural translation for nine languages.



How Neural Machine Translation can improve the work of translators

In the 2020 ATA Annual Conference in Palm Springs, there was a board. The board contained one question and the attendees were encouraged to answer the question: "Do you use machine translation (MT), and how?"

Here are some comments and direct quotes:

  • "To quickly get the gist of a paragraph. As a quick dictionary when I forget the target word in my native language."
  • "Out of curiosity! The robots are coming whether we like it or not. We should know what MT is good or not good for."
  • "To get a good laugh."
  • "To show how it does NOT work and can never replace the high quality professional work I do."
  • "Integrated in my computer assisted translation tool to assemble segments."
  • "Never for any serious kind of translation work! For laughs - to get the gist of some Facebook posts and the likes only!"
  • "To show how algorithms can't make the mental leaps that humans can!"
  • "Only when requested by the client."
  • "Then educate the client on MT, computer-assisted translation, and the difference between the two... and on human translation."

It's curious to see that most of the responses are still on the negative side, that is, they come from that angle that machine translation can be used only reactively: the translator reacts to the suggestion of the MT engine. One of the reasons for this approach is that we still think that machine translation is the same as post-editing. Also, when we think about machine translation output, we usually think of the SMT models that work like a Lego that try to fill the empty segment with the missing piece. With neural machine translation, this is not what is happening. NMT first learns the design of the puzzle, then it learns which pieces form that design. The other problem is that we have some misconceptions about Artificial Intelligence and Neural Translation and we think that it competes with our brains' functions. The human brain is much more complicated than any of the machines invented by humans. We only know its biological and physiological functions and have some ideas about the elusive realms.

What Artificial Intelligence does is to process a large amount of data and makes predictions exclusively on the basis of that data. So far we have not reached the level where artificial intelligence can use reason, judgment, or create a strategy. If it is true that AI is modeled on the human brain, only a small part of the brain was included in the study.

We also need to take a closer look at what we mean by the translation process and what is post-editing. For example, if the translator receives fragmented suggestions from the MT, then the translator is actually generating a translation by choosing to accept or reject the suggestions that are being presented to them by MT.

"This is a high-cognitive load because the translator's thoughts are constantly interrupted by the machine. On the other hand, post-editing implies that the suggestions by the MT are full suggestions that are good enough to read them."

The next question is, and I believe that the coming decade will focus on this, whether translators can be in the driver's seat using machine translation as they use TMs, TBs, or dictionaries. There are more ways to use MT than just for post-editing. We just need to have some patience to wait for the next wave. The same way we did with CAT tools and other tools and environments that - in spite of the misconception - rather than competing or replacing them, meant to help and support the translators' jobs.

Works consulted:



Comments

Popular posts from this blog

A Closer Look at Netflix's Timed Text Style Guides and Subtitling Best Practices

  Table of Contents Introduction Netflix Timed Text Style Guides Technical aspect Linguistic aspect Forced Narratives Trailers Subtitles vs. CC Conclusion Resources Appendix: SDH Identifiers Table - HU Watch my short "hook" about this post here: Download the slides  here . Introduction Subtitling and audiovisual translation Dubbing and subtitling are very creative processes. Whether the audience watches with dubbed audio, or in the original language with foreign language subtitles, closed captions, or forced narratives, the ultimate goal is to make the shows enjoyable and resourceful. As well as making sure that any text is timed appropriately to the action, capturing creative vision and nuances in translation is critical for this goal. Audiovisual translation is like creating 3D translations. In traditional translation projects, you have the source text and the target text. It's two-dimensional. With audiovisual translations, you have the source text, the visuals...

Contentful Headless CMS - l10n & i18n

Exploring Contentful for Translation and Localization It seems there is a new buzzword in the website building industry: headless CMS (Content Management System). But what is exactly a headless CMS and how does it work? In this write-up, I am going to walk you through the different steps of how I explored one of the most popular headless CMSs, Contentful; how I created a simple website with it; and what approaches I took to localize the website into another language. In the end, I realized that I needed a completely different mindset on website building and localization in general. And I had a clear vision of when a headless CMS is useful and when it is recommended to use a traditional CMS, like WordPress, instead. Finally, I learned a lot about Node.js, Gatbsy.js, website deployment, and the many challenges i18n problems introduced into the subject. Part 1: Headless CMS & Contentful What is headless CMS? A headless CMS is a platform that has no default front-end system to determin...

Time, Cost, and Quality in Localization Processes

Quality, Cost, Time dilemma in Localization Translation projects  have to meet three main requirements as defined in the ‘iron triangle’ invented by Dr. Martin Barnes in the 1970s: quality, time, and cost. One side of the triangle cannot be changed without affecting the other two. According to this theory, translation projects appear to be locked in an impossible equation where the ideal process that is quick, well done, and cheap can never be achieved. According to this idea, projects can be: Fast and cheap  = this results in poor quality. This is what most clients are asking for, without taking into consideration that reducing the amount of time spent on a project will most probably cause issues. Fast and high quality  = expensive. You will need a bigger team, which will cost you more in terms of human resources and organization. High quality and cheap  = slow. In order to save money, the translation agency will choose beginners and give them extra time, and t...