情侣自拍

Skip to content
  • Shop ATAware
  • Contact Us
  • Log In Welcome,
情侣自拍 (ATA)
Find a Translator or Interpreter
  • Client Assistance
    • Find a Translator Button
      • Find a Language Professional
    • Client Resources
      • Why Should I Hire a Professional?
      • Translator vs. Interpreter
      • Buying Language Services
      • What is Machine Translation?
    • More Client Resources
      • Why Hire an ATA-Certified Translator?
      • Need a Certified Translation?
      • The ATA Compass Blog
      • Know Your Rights to Language Access
  • Certification
    • Register Buttons
      • Order Practice Test

      • Register for Exam
    • Client Resources
      • Why Hire an ATA-Certified Translator?
      • What is a Certified Translation?
    • About the Exam
      • How to Prepare
      • Practice Test
      • How the Exam is Graded
      • Exam Schedule
      • Need More Information?
    • Already Certified?
      • Put Your Credentials To Work
      • Continuing Education Requirement
  • Career and Education
    • Event Buttons
      • Visit ATA67

      • Upcoming Webinars
    • For Newcomers
      • Student Resources
      • Starting Your Career
      • The Savvy Newcomer Blog
    • For Professionals
      • Growing Your Career
      • Business Strategies
      • Next Level Blog
      • Client Outreach Kit
      • Mentoring Opportunities
    • More Resources
      • Educators and Trainers
      • Tools and Technology
      • Publications
      • School Outreach
  • Events
    • Event Buttons
      • Visit ATA67

      • Upcoming Webinars
        听
    • Events
      • Annual Conference
      • Free Events for ATA Members
      • Certification Exam Schedule
    • More Events
      • Virtual Workshops and Events
      • Live and On-Demand Webinars
      • Calendar of Events
  • News
    • Industry News
    • Advocacy and Outreach
    • The ATA Chronicle
    • The ATA Podcast
    • ATA Newsbriefs
    • Press Releases
  • Member Center
    • Member Buttons
      • Join ATA

      • Renew Now
    • Member Resources
      • Join ATA
      • Renew Your Membership
      • Benefits of Membership
      • Divisions & Special Interest Groups
      • Chapters, Affiliates, and Other Groups
      • Get Involved
      • Member Discounts
      • Shop ATAware
    • Already a Member?
      • Connect with Members
      • Credentialed Interpreter Designation
      • Become a Voting Member
      • Submit Member News
      • Submit Your Event
      • Contact Us
  • About Us
    • Contact Button
      • Contact ATA

    • About ATA
      • Who We Are
      • Honors and Awards Program
      • Advertise with Us
      • Media Kit
    • How ATA Works
      • Board of Directors
      • Committees
      • Policies & Procedures
      • Code of Ethics
      • ATA Team
  • Join ATA
  • Renew Your Membership
  • Contact Us
  • Find a Translator or Interpreter
  • Search for:
September 9, 2016

Neural Machine Translation

Resources, Tools and Technology
Source: The ATA Chronicle

Do we need to start shivering in fear when we hear folks talking about neural machine translation?

There has been much in the news lately about the next wave of machine translation (MT) technology driven by something called deep neural nets (DNN). With the help of some folks who understand more about it than I do, I鈥檝e attempted to provide a brief overview about what this is. First, I need to confess that I will be trying to explain something here that I don鈥檛 fully understand myself. Still, I hope that my research has helped me comprehend and communicate some basic underlying principles.

Most lectures you鈥檝e listened to about MT in the past few years have likely included a statement like this: 鈥淭here are basically two kinds of MT鈥攔ule-based and statistical MT鈥攁nd a third that combines the two鈥攈ybrid MT.鈥 You鈥檝e also probably heard that rule-based MT was the earliest form of MT in the computer age, going back all the way to the 1950s. Back then, this form of MT consisted of a set of rules about the source and target language and included a dictionary. The transfer between the source and target language in rule-based MT happens either via an 鈥渋nterlingua,鈥 a computerized representation of the source text, or directly between the source and target language.

Statistical machine translation (SMT), on the other hand, became all the rage in the early 2000s. (The first commercial offering, LanguageWeaver [now owned by SDL], was launched in 2002; the widely used open-source engine Moses emerged in 2005; Google and Microsoft switched to statistical MT in 2007; and Yandex and Baidu started using SMT as recently as 2011.) SMT, or more accurately for all of these implementations, 鈥減hrase-based statistical machine translation,鈥 is trained on bilingual data and monolingual data. It parses the data into 鈥渘-grams,鈥 which are phrases consisting of an 鈥渘鈥 number of words. The same thing happens to the source segment in the translation process. The source n-grams are then matched with target n-grams, which are then combined to form whole segments again鈥攁nd that鈥檚 often where things go awry. (This is why SMT can prove to be a much richer resource when using an approach that just looks for fragments rather than whole segments.) Another potential flaw with SMT is the faulty selection process when the system tries to decide which of the many possible target n-grams to use. One way to guard against bad choices is by validating them on the basis of the monolingual target data on which the system was trained, but that only goes so far. (By the way, that鈥檚 why an approach that offers access to more than just one of those n-gram fragments at a time within a translation environment tool has to be one of the up-and-coming developments.)

Neural machine translation (NMT)鈥攁nd let鈥檚 pause and be thankful that one of this technology鈥檚 first proposed terms, 鈥渞ecursive hetero-associative memories for translation鈥 (coined by Mikel L. Forcada and Ramon P. Neco in 1997) did not survive鈥攊s an extremely computing-power-heavy process (which is why it didn鈥檛 go anywhere in 1997).1 It鈥檚 part of the larger field of 鈥渕achine learning.鈥 In 1959, Arthur Samuel, a pioneer in the field of artificial intelligence and machine learning, defined machine learning as the 鈥渇ield of study that gives computers the ability to learn without being explicitly programmed.鈥2

In SMT, the focus is on translated phrases that the computer is taught, which are then reused and fitted together according to statistics. NMT, on the other hand, uses neural networks that consist of many nodes (conceptually modeled after the human brain), which relate to each other and can hold single words, phrases, or any other segment. These nodes build relationships with each other based on bilingual texts with which you train the system. Because of these manifold and detailed relationships, it鈥檚 possible to look at not just limited n-grams (as in SMT), but at whole segments or even beyond individual segments. This allows for the formation of significantly more educated guesses about the context, and therefore the meaning, of any word in a segment that needs to be translated. For instance, it鈥檚 at least theoretically unlikely to have 鈥淧rince鈥 translated as a (royal) prince by the NMT in a sentence like 鈥淭he music world mourns the death of Prince,鈥 as Google, Microsoft, Yandex, and Baidu all do at the moment. (By the way, I鈥檓 mourning as well.)

In languages like German with separable verbs, such as umfahren (鈥渞un over鈥), there is a much greater likelihood that the system will notice the missing part of the verb at the end of the sentence if the machine doesn鈥檛 have to bother with chopping it into n-grams first. Take, for example, the simple sentence, Ich fahre den Fu脽g盲nger um (鈥淚 run over the pedestrian鈥). Bing translates it (today) as 鈥淚鈥檓 going to the pedestrian,鈥 and Google renders it as 鈥淚 drive around the pedestrian.鈥 Only Yandex gets it right. (Baidu does not offer this language combination.)

Machine learning (itself a subfield of artificial intelligence) also comes into play as common usage gradually forges certain linguistic connections (e.g., 鈥渕usic world鈥 and 鈥淧rince鈥; 鈥渇ahren鈥 and 鈥渦m鈥). This means that, just as Arthur Samuel predicted, the computer continues to 鈥渓earn鈥 without explicitly being programmed.

At least theoretically, the NMT approach is very promising for generic engines like those of the search engines mentioned above (Google, Microsoft, Yandex, and Baidu). This is because 鈥渃ontext鈥 does not necessarily have to be specified by the training data, but can be recognized by the system evaluating the context (provided that the user supplies more than just a word or single phrase). So, you won鈥檛 be surprised to hear that all those companies have already entered the realm of NMT. Naturally they don鈥檛 reveal how much of their present system is 鈥渘eural鈥 versus 鈥渟tatistical only,鈥 but chances are it鈥檚 a mix of both. And that would make all the more sense since one of the ways to use NMT is in combination with SMT鈥攅ither as a quasi-independent verification process or as an integrated process that helps in selecting the 鈥渞ight鈥 n-grams.

In some areas similar processes have already demonstrated remarkable success, including some that are used by search engines such as Google Image Search (which can be used very effectively for cross-language searches and be helpful in the translation process).

You probably read that Facebook launched its own MT system earlier this year specifically geared for the very casual language of its users. While that system is still mostly SMT-based, Facebook is working on an NMT solution as well. You might want to take a look at a presentation by Alan Packer, Facebook鈥檚 director of engineering and language technology (formerly of Microsoft), entitled 鈥淯nderstanding the Language of Facebook.鈥3

One misconception in Packer鈥檚 presentation is his description of all this as a linear development. He paints SMT as more or less having run its course, now to be taken over by NMT. While I understand that someone so deeply embedded in one particular field must automatically think it the only worthwhile one, it鈥檚 really unlikely to be the case. The same was said in the early days about rule-based MT (RbMT) by proponents of SMT, and that assumption has not proven to be true. Many systems are using a hybrid approach between SMT and RbMT, and for some language combinations RbMT might still be a better solution (especially for language pairs that are very close to each other, like Catalan and Spanish or Croatian and Serbian).

But are we on the verge of a big new breakthrough overall? To answer that, you might want to look through the joint presentation by Diego Bartolome (Tauyou Language Technology) and Gema Ramirez (Prompsit Language Engineering), 鈥淏eyond the Hype of Neural Machine Translation.鈥4 Since there is no open-source toolkit for NMT, like Moses for SMT, very few companies actually offer customized NMT systems. There are components like the deep learning frameworks Theano and Torch and specific NMT software like GroundHog and seq2seq, but these are anything but user-friendly and require significant expertise. Using them to build the NMT engine takes a lot of computing power (10 CPUs or 1 GPU鈥攇raphics processing unit) and time (about two weeks of training per language pair once the training data is assembled and cleaned). Tauyou Language Technology and Prompsit Language Engineering are some of the first vendors who are working on commercial versions of NMT. (Interestingly, Tauyou comes with an SMT background, and Prompsit with a background in RbMT). While they are not actively selling the NMT solutions yet, they are doing a lot of pilots, as you鈥檒l see from the presentation. The results of these pilots are mixed.

I already mentioned the much larger processing and time requirements. There are also limitations as far as the number of words per language that can be trained with the processing power currently available to mere mortals (in opposition to companies like Google), the approximately three-fold time the system takes to actually translate, and the fact that retraining the system with new data would once again take two weeks. But there are some improvements in the quality鈥攁lthough, according to the presentation, this is not adequately appreciated by translators. (I assume this has to do with even less predictability when it comes to post-editing鈥攁nd presumably even more erratic decisions when it comes to partial suggestions.) However, this is still very early in the game, so I wouldn鈥檛 be surprised to see the quality continue to improve.

So, do we need to start shivering in fear when we hear folks talking about NMT? Although I don鈥檛 completely understand the technology, I (and now you) have seen numbers showing only moderate progress. So, no, we鈥檒l continue to be very assured of our jobs for a long time. I do look forward, though, to seeing how NMT will creatively find its way into our translation environment and improve our work.

Notes
  1. Forcada, Mike L., and Ramon P. Neco. 鈥淩ecursive Hetero-Associative Memories for Translation,鈥 In Biological and Artificial Computation: From Neuroscience to Technology (Springer, 1997), 453鈥462, .
  2. Samuel, Arthur, 鈥淪ome Studies in Machine Learning Using the Game of Checkers,鈥 IBM Journal (July 1959), .
  3. Packer, Alan. 鈥淯nderstanding the Language of Facebook,鈥 .
  4. Bartolome, Diego, and Gema Ramirez. 鈥淏eyond the Hype of Neural Machine Translation,鈥 MIT Technology Review
    (May 23, 2016), .

Jost Zetzsche is the co-author of Found in Translation: How Language Shapes Our Lives and Transforms the World, a robust source for replenishing your arsenal of information about how human translation and machine translation each play an important part in the broader world of translation. Contact: jzetzsche@internationalwriters.com.

This column has two goals: to inform the community about technological advances and at the same time encourage the use and appreciation of technology among translation professionals.

Share this

Posts navigation

← ATA 2016 Elections: Candidate Statements
Curiosity as an Approach to Marketing →

Latest Posts

  • Coming Soon: ATA Microcredential Series May 4, 2026
  • Introducing the ATA Learning Hub! May 4, 2026
  • Member News May 4, 2026
  • Texas Court Interpreter Detained by ICE at Airport Says She鈥檚 Been 鈥楬umiliated and Treated Like a Criminal鈥 May 4, 2026
  • U.S. Department of Education Dissolving Federal Office Serving English Learners May 4, 2026
  • Northern Ireland to Offer Free Sign Language Classes for Deaf Children May 4, 2026
  • Washington State Passes Law to Promote Consistent Language Access May 4, 2026
  • Breaking News: Texas Interpreter Meenu Batra Released from ICE Custody May 4, 2026
  • Newsbriefs: April 30, 2026 May 1, 2026
  • Stand with Meenu Batra: A Call to Action for Language Professionals April 21, 2026

Topics

  • Advocacy & Outreach
  • Annual Conference
  • Book Reviews
  • Business Strategies
  • Certification Exam
  • Certification Program
  • Client Assistance
  • Educators and Trainers
  • Growing Your Career
  • Industry News
  • Interpreting
  • Member Benefits
  • Member News
  • Mentoring
  • Networking
  • Public Outreach
  • Publications
  • Resources
  • School Outreach
  • Specializations
  • Starting Your Career
  • Student Resources
  • Tools and Technology
  • Translation
Find a Translator听 or Interpreter
ata_logo_footer

情侣自拍
211 N. Union Street, Suite 100
Alexandria, VA 22314

Phone +1-703-683-6100
Fax +1-703-778-7222

  • Certification
  • Career and Education
  • Client Assistance
  • Events
  • News
  • Member Center
  • About Us
  • Contact Us
  • Sitemap
  • Privacy Policy
  • Accessibility Statement
  • Submit Feedback

漏 2026 -听情侣自拍

Find a Translator or Interpreter
Scroll To Top
By clicking accept or closing this message and continuing to use this site, you agree to our use of cookies.