Machine Translation. We can do better!

机器翻译。我们可以做得更好!

2022-06-30 22:50 TAUS

本文共1231个字,阅读需13分钟

阅读模式 切换至中文

MT has come a long way. After seventy years of research, the technology is now taken into production. And yet, we are missing out on the full opportunities. Because the developers are preoccupied with the idea that the massive models will magically solve the remaining problems. And because the operators in the translation industry are slow in developing new MT-centric translation strategies. This article is an appeal to everyone involved in the translation ecosystem to come off the fence and realize the full benefits of MT. We can do better! From grassroots to MT industrialization Gartner, in their AI Landscape 2020 blog article, declared industrialization and democratization of AI the two dominant trends for AI in 2021. All the hard work and experimentation with models and data by the early adopters and MT gurus has finally paid off. We are shifting from a bottom-up grassroots movement to a top-down directive coming from the executive suites. Enterprise-wide adoption of MT becomes part of the AI or digital transformation program and belongs now to the responsibility of the CIO or CTO. The technology is there to simply translate everything, launch new platforms and talk to many more users in their own languages. But then, too often we realize that the quality of automated translation is not on par with enterprise production requirements in terms of reliability and trust. The technology may be there, but the delivery is not yet good enough. We can do better! If we set the right priorities in our strategies. The big-tech MT developers Most MT is sourced from the big tech companies Amazon, Google and Microsoft. They have the scale and the capital to develop the massive models that return the amazing results that we have seen in the last few years and that have been the driving force behind the industrialization of MT. The rapid improvements have been feeding an unprecedented optimism on the West Coast of the United States that remaining issues such as MT hallucinations, catastrophic errors and domains and languages not being covered yet by the technology will all go away within five to ten years. One disturbing factor though is that the massive MT models are black boxes. Even the researchers who train them can’t tell exactly why one performs better than the other. The model work is glamorous and cool, but the intellectual insight that would allow us to reproduce bugs and remove them is hard to get. To get models to work in production, more data engineering is required than research. Well-focused data engineering can bring in the nuances that are required for robust performance in a real-world domain. The problem is that most researchers like to do the model work, not the data work. (See this Google Research article: Data Cascades in High-Stakes AI). Many MT platforms provide customization features allowing users to upload translation data and take care of their own data engineering work. These features however, as TAUS found out, require a lot of experimentation and experience.* In-domain training data have unpredictable, often low and sometimes even negative impact on the performance of the engines. It seems that the big tech companies treat their customization features as stop-gap measures for the time it takes until human parity is reached. Five to ten years? The big tech MT developers can do better to support and facilitate the industrialization of MT. Here is how: Don’t bet the future entirely on the brute force of the massive models Improve your customization features to better support your business customers in building production-ready engines. MT users The technology breakthroughs in the past years have caused a rise in the adoption of MT. Nothing spectacular or revolutionary though. The MT engines are simply plugged into the existing workflows and are being used as complementary sources for translation matches. Translators see their tasks changing more and more to post-editing. The new technology is mainly used to help the business drive for continuous efficiency gains and lower word rates, very much so in the tradition of thirty years of leveraging translation memories. What we miss in the translation industry overall is blue-sky thinking. Putting aside a few start-up innovators, most of the actors in the translation industry have taken a defensive approach towards MT technology. The result is a general negative sentiment with emphasis on cost reductions, compromises in translation quality, disruption in the workforce and pessimistic perspectives on the future of the industry. The problem is that we are all so deeply rooted in our traditions, we can’t see through the present. In their Market Guide for AI-Enabled Translation Services (June 2022) Gartner recommends that companies divide content into “tiers” of “acceptable translation quality” and develop new end-to-end workflows taking into account automation enabled by MT technology. Some start-up innovators have done exactly that, by putting MT technology at the core of a brand new real-time multilingual business solution. MT technology can be a force multiplier for those operators in the translation industry that are capable of shifting from a defensive to a proactive approach. MT users, LSPs and enterprises can do better to support and facilitate the industrialization of MT. Here is how: Focus on data engineering. Do not accept that the quality output of, among others, the Amazon, Google, Microsoft and Systran engines is as good as it can get. Significant improvements can be made using core competencies such as domain knowledge and linguistic expertise. Design end-to-end MT-centric workflows. Do not think of MT as just an add-on to your current process and workflow but make it the core of new solutions serving new customers, translating content that was never translated before. Provide new opportunities for linguists. Post-editing is not the end-game. Create new perspectives by leveraging intellectual insights for better automation. TAUS recipe for better MT TAUS has been an industry advocate for translation automation since 2005. We have developed a unique recipe for better MT as outlined here below. The first step in every MT project is to measure and evaluate the translation quality. Most MT users are just measuring and comparing the baseline engines. TAUS takes the evaluation a step further. We train and customize different MT engines and then select the engine with the maximum achievable quality in the customer domain. See TAUS DeMT™ Evaluate. The second step is the creation of in-domain customer-specific training datasets, using a context-based ranking technique. Language data are sourced from the TAUS Data Marketplace, from the customer’s repositories or created on the Human Language Project platform. Advanced automatic cleaning features are applied. See TAUS DeMT™ Build. The third step is then generating the improved machine translation. Improvements demonstrated show scores between 11% and 25% over the baseline engines from Amazon, Google and Microsoft. In many cases, this brings the quality up to levels equal to human translation or post-edited MT. Some customers refer to DeMT™ Translate as ‘zero-shot localization’, meaning that translated content goes directly to customers without post-editing. TAUS offers DeMT™ Translate via an API to LSPs and enterprises as a white-label product. * MT customization features require a lot of experimentation and experience. See TAUS DeMT™ Evaluation Report and contact a TAUS expert to learn how to best work with MT customization.
MT已经走过了漫长的道路。经过70年的研究,这项技术现已投入生产。然而,我们却错过了所有的机会。因为开发者们一心想着大规模的模型会神奇地解决剩下的问题。而且因为翻译行业的运营商在开发新的以MT为中心的翻译策略方面进展缓慢。本文旨在呼吁翻译生态系统中的所有参与者摆脱藩篱,实现MT的全部优势。我们可以做得更好! 从草根到MT产业化 Gartner在其《AI Landscape 2020》博客文章中宣布,AI的工业化和民主化是2021年AI的两大主导趋势。早期采用者和MT专家对模型和数据的所有辛勤工作和实验终于得到了回报。我们正在从自下而上的基层运动转变为来自行政部门的自上而下的指令。在整个企业范围内采用机器翻译成为人工智能或数字化转型计划的一部分,现在属于首席信息官或首席技术官的责任。这项技术可以简单地翻译所有内容,推出新的平台,并用更多用户的语言与他们交谈。但是,我们常常意识到,自动化翻译的质量在可靠性和可信度方面无法满足企业的生产要求。技术可能已经存在,但交付还不够好。我们可以做得更好!如果我们在战略中设定了正确的优先顺序。 高科技MT开发商 大多数移动终端都来自大型科技公司亚马逊、谷歌和微软。他们有足够的规模和资金来开发大型模型,这些模型在过去几年里取得了令人惊叹的成果,这也是MT产业化背后的推动力。这些快速的进步让美国西海岸的人们产生了前所未有的乐观情绪,他们认为,诸如MT幻觉、灾难性错误以及该技术尚未覆盖的领域和语言等遗留问题将在5到10年内全部消失。 然而,一个令人不安的因素是,大规模的MT模型是黑盒子。即使是训练他们的研究人员也不能确切地说出为什么一个人比另一个人表现得更好。模型的工作是迷人的和酷的,但智力的洞察力,将使我们能够复制错误,并删除它们是很难得到的。为了让模型在生产中发挥作用,需要更多的数据工程而不是研究。重点突出的数据工程可以带来在现实领域中实现稳健性能所需的细微差别。问题是,大多数研究人员喜欢做模型工作,而不是数据工作。(See这篇Google研究文章:高风险AI中的数据级联)。 许多MT平台提供自定义功能,允许用户上传翻译数据并处理自己的数据工程工作。然而,TAUS发现,这些功能需要大量的实验和经验。域内训练数据对引擎的性能具有不可预测的、通常较低的、有时甚至是负面的影响。大型科技公司似乎把他们的定制功能视为权宜之计,直到人类达到平等。五到十年? 大型科技机器翻译开发商可以更好地支持和促进机器翻译的产业化。具体操作如下: 不要把未来完全押在大规模模型的蛮力上 改进您的自定义功能,以更好地支持您的业务客户构建生产就绪型引擎。 MT用户 过去几年的技术突破使MT的采用率上升。不过没什么特别的或革命性的。MT引擎只需插入现有工作流程,即可用作翻译匹配的补充来源。翻译人员发现他们的任务越来越多地变成了后期编辑。这项新技术主要用于帮助企业不断提高效率和降低字数,这与三十年来利用翻译记忆库的传统非常相似。 我们在翻译行业中所缺少的是一种空穴来风的思维方式。除了少数刚起步的创新者之外,翻译行业的大多数参与者都对MT技术采取了防御性的做法。其结果是普遍的负面情绪,强调降低成本、牺牲翻译质量、破坏劳动力队伍,并对行业的未来持悲观态度。问题是我们都深深地植根于我们的传统,我们不能看透现在。 Gartner在其《人工智能翻译服务市场指南》(2022年6月)中建议,公司应将内容划分为"可接受的翻译质量"的"层",并开发新的端到端工作流,同时考虑MT技术实现的自动化。一些初创的创新者正是这样做的,他们将MT技术置于全新的实时多语言业务解决方案的核心。 对于翻译行业中能够从防御性方法转变为主动性方法的运营商而言,MT技术可以成为一种力量倍增器。 MT用户、LSP和企业可以更好地支持和促进MT的产业化。具体操作如下: 专注于数据工程。不要接受亚马逊、谷歌、微软和Systran引擎等的质量输出是最好的。使用核心能力(如领域知识和语言专业知识)可以实现显著的改进。 设计以MT为中心的端到端工作流。不要将机器翻译仅仅看作是您当前流程和工作流的一个附加组件,而要将其作为服务新客户的新解决方案的核心,翻译以前从未翻译过的内容。 为语言学家提供新的机会。后期编辑并不是最后的游戏。通过利用智能洞察力创建新的视角,实现更好的自动化。 TAUS是更好的MT的秘诀 自2005年以来,TAUS一直是翻译自动化的行业倡导者。我们已经开发出一个独特的配方,更好的MT如下所述。 每一个机器翻译项目的第一步都是衡量和评估翻译质量。大多数MT用户只是测量和比较基准发动机。TAUS将评估进一步推进。我们培训和定制不同的MT发动机,然后在客户领域选择具有最高可实现质量的发动机。参见TAUS DeMT™评价。 第二步是使用基于上下文的排名技术,创建领域内客户特定的训练数据集。语言数据来源于TAUS数据市场、客户的存储库或在人类语言项目平台上创建。应用高级自动清洁功能。参见TAUS DeMT™构建。 第三步是生成改进的机器翻译。所展示的改进显示,与亚马逊、谷歌和微软的基准引擎相比,得分在11%到25%之间。在许多情况下,这会使质量达到人工翻译或后期编辑MT的水平。一些客户将DeMT™ Translate称为“零拍本地化”,这意味着翻译后的内容无需后期编辑即可直接提供给客户。TAUS通过API向LSP和企业提供DeMT™ Translate作为白标产品。 * MT定制功能需要大量的实验和经验。请参阅TAUS DeMT™评估报告并联系TAUS专家,了解如何最好地使用MT定制。

以上中文文本为机器翻译,存在不同程度偏差和错误,请理解并参考英文原文阅读。

阅读原文