What ChatGPT Gets Right and Wrong and Why It’s Probably a Game-changer for the Localization Industry

ChatGPT的对错以及为什么它可能会改变本地化行业的游戏规则

2023-01-17 14:01 lionbridge

本文共1492个字,阅读需15分钟

阅读模式 切换至中文

The explosion of ChatGPT into the mainstream since its launch on November 30, 2022, has generated unprecedented attention and commentary. After spending a few days (and nights) having conversations with ChatGPT, these are the only questions that I think matter: What does it get right? What does it get wrong? How can we use it? Want to learn the answers to these questions and what I foresee for this new localization tool? Read on. We’ll also post a longer, more detailed whitepaper on this topic soon. What Does ChatGPT Get Wrong? There are things that you can’t rely on ChatGPT for: It doesn’t tell the truth. ChatGPT says things people want to hear, given the information it has. It doesn’t have a clue about the real world. It only knows a curated version of what people say about the real world, using that information to create very convincing language to represent what it’s learned. It can’t count. I tried replicating Jonas Degrave’s simulation at Engraved using a more complicated calculation, and got an erroneous result. It does correctly attempt to multiply the two numbers from this Python command it knows how to simulate, which is quite remarkable. However, it just can’t count. It can’t think. It will often write something that outlines accurate premises and true statements about what I asked it to do, and confidently apply it erroneously. Basically, it can’t reason; it’s not a finite state machine. Its humble bragging is a little cringe. ChatGPT will write with deep aplomb and authority, yet sheepishly tells you it’s only here to help, and exercises embarrassing contriteness when you tell it that it got it wrong. What ChatGPT Gets Right Having gotten through what ChatGPT isn’t or can’t do, let’s look at what it can do as a text interpreter and generator. ChatGPT can: Write better than you. I developed this strong belief after having many conversations with ChatGPT. It writes at various levels of complexity with a wide range of vocabulary. This generative AI is as good, in my opinion, as the top decile of human content writers. Follow instructions. It will flawlessly change texts in specific ways, both in form and content. Most crucially, it maintains the conversation context and understands when you refer to what it had done before, even in the most vernacular manner. Modify text while keeping meaning. Given any text, and following instructions, ChatGPT makes any modification to the content, form, and style. It maintains the semantic content of the text or modifies it as you ask it to. Manage multilingual terminology, a critical issue in localization. I don’t know yet if it’s realistic for large-scale translation. But, I could see ChatGPT doing a decent job of introducing specific terminology while editing previously translated material, even if it’s not the original translator. Detect offensive text. I provided ChatGPT excerpts from a pleading in a federal criminal case with racist and homophobic text messages, then asked it to identify offensive text. It did a great job and was able to explain its reasoning. Perform entity detection. I asked ChatGPT to perform a typical case of entity detection and to place tags around the entities. It missed a couple. But, with an additional couple of prompts, tagged them up easily. Classify things to a taxonomy. One of the most mind-boggling things about ChatGPT is how it can apply general knowledge to a specific situation. We’ve established that you can’t rely on ChatGPT to say true things or to know what’s correct. (This means content creators would have to check whether it’s talking nonsense.) However, once we have a text containing meaning that we’re happy with, ChatGPT can manipulate or transform its form and content -- while retaining encoded meaning. This is an opportunity for localization providers because we don’t have to generate meaningful content from scratch. Let’s take a closer look at the localization activity landscape and how ChatGPT could affect what we’re doing today. This generative AI: Is very skilled at translation. For languages with large corpora, ChatGPT is likely on par with state-of-the-art engines for out-of-the-box Machine Translation output, if not superior in some respects. Seems quite effective at following terminological instructions. Can apply style instructions, either broadly or narrowly. In contrast, getting Machine Translation engines to perform these types of tasks correctly is challenging. Is quite good at categorizing things, particularly for text, to arbitrary taxonomies. This ability is beneficial for the localization industry, as we may want to apply specific instructions to certain types of content and other instructions to others. Excels at editing given text, the most crucial element of quality localization. It’s quite good at the four major components of reviewing and translating text. Helps with content analysis, analyzing text for effective processing, improvement, or ROI. It can help anticipate or preempt translation quality issues, tune it for outcomes effectiveness like reach, SEO, CTA/CTR performance, and improve legibility at source and in the target, among others. Helps with writing and editing working code. Elite programmers debate whether ChatGPT writes at their level, and it appears not. However, I asked it to write some XML content extraction code, and it produced working code that I could run. It makes creating code (and learning how) easier for non-coders. ChatGPT Will Specify New Skills and Practices I’ve developed the beginning of an understanding of what type of prompting ChatGPT requires. This special technology only takes natural language as input. To use this technology in production, we’ll need to develop expertise for effective prompting. Specific transformations of content will likely require successions of prompts that will each perform different tasks: cleanup, pre-and post-processing, etc. Learning to use natural language prompts as part of our automation pipelines, in a way that is both contextually relevant and sufficiently predictable in the output, will be an interesting journey. Where Do We Go from Here? We now definitively know we can’t ignore this new generative AI. It’s likely to disrupt our industry. So, we must lead and drive that push to language automation, lest we get left behind. ChatGPT can transform and annotate text on par with the average human editor and likely perform these tasks more efficiently. It can perform tasks relying on a diversity of skills virtually no individual human possesses, and it can generalize its knowledge to new situations. Most importantly, it shows potential for solving some lingering localization automation problems. Of course, it’s one thing to have a conversation with it using toy examples. It’s another thing to imagine using it at scale to perform these actions. Moving forward, we must: Conduct real-world tests at scale to evaluate error rates for each type of localization and editing task investigated here. Analyze detailed macro and micro user journeys occurring within the localization value chains and identify where they will likely be disrupted with this type of text automation. Understand how to prompt and provide relevant context to ChatGPT at scale, and document pitfalls and best practices. Develop the new automation and human-in-the-loop editing workflows, inventing what post-editing and QA will mean tomorrow with such an AI in the loop. Design new automation and User Experience (UX) interaction contexts for both localization agents and customers for each possible improvement opportunity. Ensure the economics of licenses, deployment costs, and maintenance makes sense for our business Some Thoughts About Language and Real-World Usage of Chat GPT One of the things that I found most striking was how ChatGPT got complex number operations almost right, but still wrong. It doesn’t cheat. It really learns everything from the language it trains on. The fact that it finds almost the correct result of an operation beyond a certain order of magnitude (and the correct result for smaller numbers) tells us that a language corpus of a sufficient scale contains statistically significant knowledge about the real world. But, it also shows that dedicated formal systems (such as mathematics) are required to produce meaningful, reliable, and accurate information about the real world. ChatGPT offers a sobering reminder that a self-referential, self-consistent system cannot in and of itself carry the truth of the world, which exists independently. This echoes Gödel’s incompleteness theorem. As conscious beings, we can’t untether our cognition from formal and material systems grounding our understanding of the world in a reality that imposes itself on us and that we cannot define away through language alone. Contact us Have your own translation or localization project? Need to ensure it gets done accurately, quickly, and under budget? We’ll use innovative generative AI like ChatGPT to help. Contact us today to find out more about Lionbridge’s translation and localization services.
自2022年11月30日推出以来,ChatGPT成为主流,引起了前所未有的关注和评论。在花了几天 (和晚上) 与ChatGPT进行对话之后,我认为这些是唯一重要的问题: 什么是正确的? 它出了什么问题? 我们如何使用它? 想了解这些问题的答案,以及我对这个新的本地化工具的预见?继续阅读。我们还将很快发布有关此主题的更长,更详细的白皮书。 ChatGPT错了什么? 有些事情你不能依赖ChatGPT: 它没有说实话。ChatGPT说人们想听的东西,因为它有信息。 它没有关于现实世界的线索。它只知道人们对现实世界的看法的精选版本,使用这些信息来创建非常有说服力的语言来代表所学到的内容。 这不算。我尝试使用更复杂的计算来复制Jonas Degrave的模拟,结果得到了错误的结果。它确实正确地尝试从这个Python命令中乘以两个数字,它知道如何模拟,这是相当了不起的。但是,它无法计数。 无法思考。它通常会写一些概述我要求它做的准确前提和真实陈述的东西,并自信地错误地应用它。基本上,它不能推理; 它不是有限状态机。 它卑微的吹牛有点畏缩。ChatGPT将以深切的沉着和权威来写作,但令人毛骨悚然地告诉您,这只是在这里提供帮助,当您告诉它弄错了时,您会感到尴尬。 ChatGPT做对了什么 了解了ChatGPT不是或不能做的事情后,让我们看看它作为文本解释器和生成器可以做什么。ChatGPT可以: 写得比你好。在与ChatGPT进行了多次交谈后,我建立了这种坚定的信念。它以各种复杂程度写作,词汇量广泛。在我看来,这种生成性人工智能和人类内容作家的前十位一样好。 遵循说明。它将以特定的方式完美地改变文本,无论是形式还是内容。最重要的是,它可以保持对话的上下文,并理解您何时提及它以前所做的事情,即使以最白话的方式也是如此。 在保留意义的同时修改文本。给定任何文本并遵循说明,ChatGPT会对内容,形式和样式进行任何修改。它维护文本的语义内容或根据您的要求对其进行修改。 管理多语言术语,这是本地化中的一个关键问题。我还不知道大规模翻译是否现实。但是,我可以看到ChatGPT在编辑以前翻译的材料时在介绍特定术语方面做得不错,即使它不是原始的翻译。 检测攻击性文本。我提供了联邦刑事案件中诉状的ChatGPT摘录,其中包含种族主义和仇视同性恋的短信,然后要求其识别令人反感的短信。它做得很好,并且能够解释其推理。 执行实体检测。我要求ChatGPT执行实体检测的典型案例,并在实体周围放置标签。它错过了一对夫妇。但是,通过另外几个提示,可以轻松标记它们。 将事物分类为分类法。关于ChatGPT最令人难以置信的事情之一是它如何将一般知识应用于特定情况。 我们已经确定,您不能依靠ChatGPT说出真实的事情或知道正确的事情。(这意味着内容创建者必须检查它是否在胡说八道。)但是,一旦我们有了包含我们满意的含义的文本,ChatGPT就可以操纵或转换其形式和内容-同时保留编码的含义。对于本地化提供商来说,这是一个机会,因为我们不必从头开始生成有意义的内容。让我们仔细看看本地化活动的情况,以及ChatGPT如何影响我们今天的工作。这个生成AI: 非常擅长翻译。对于具有大型语料库的语言,ChatGPT可能与开箱即用的机器翻译输出的最新引擎相当,即使在某些方面并不优越。 似乎在遵循术语说明方面非常有效。 可以广泛或狭义地应用样式说明。相比之下,让机器翻译引擎正确执行这些类型的任务是具有挑战性的。 非常擅长将事物 (尤其是文本) 分类为任意分类法。这种能力对本地化行业是有益的,因为我们可能希望将特定的指令应用于某些类型的内容,并将其他指令应用于其他类型的内容。 擅长编辑给定的文本,这是质量本地化的最关键要素。它在复习和翻译文本的四个主要组成部分方面相当出色。 帮助进行内容分析,分析文本以进行有效的处理,改进或ROI。它可以帮助预测或抢占翻译质量问题,调整其效果,如reach,SEO,CTA/CTR性能,并提高来源和目标的易读性等。 帮助编写和编辑工作代码。精英程序员争论ChatGPT是否在他们的级别上写作,但事实并非如此。但是,我要求它编写一些XML内容提取代码,它产生了我可以运行的工作代码。它使非编码人员更容易创建代码 (以及学习如何)。 ChatGPT将指定新的技能和实践 我已经开始了解ChatGPT需要哪种类型的提示。这种特殊的技术只以自然语言作为输入。要在生产中使用这项技术,我们需要开发专业知识以进行有效的提示。内容的特定转换可能需要连续的提示,每个提示将执行不同的任务: 清理,预处理和后处理等。学习将自然语言提示作为我们自动化管道的一部分,以一种既与上下文相关又在输出中充分可预测的方式,将是一个有趣的旅程。 我们从这里去哪里? 我们现在明确知道我们不能忽视这种新的生成性人工智能。这很可能会扰乱我们的行业。因此,我们必须领导并推动语言自动化的发展,以免我们落后。ChatGPT可以像普通的人工编辑器一样转换和注释文本,并且可能更有效地执行这些任务。它可以依靠几乎没有个人拥有的各种技能来执行任务,并且可以将其知识推广到新情况。 最重要的是,它显示出解决一些挥之不去的本地化自动化问题的潜力。当然,用玩具例子和它对话是一回事。想象一下,大规模使用它来执行这些操作是另一回事。向前迈进,我们必须: 大规模进行实际测试,以评估此处调查的每种定位和编辑任务的错误率。 分析本地化价值链中发生的详细的宏观和微观用户旅程,并确定这种类型的文本自动化可能会破坏它们的位置。 了解如何大规模提示并向ChatGPT提供相关上下文,并记录陷阱和最佳实践。 开发新的自动化和人工在环编辑工作流程,通过循环中的AI来发明后期编辑和质量检查的意义。 为每个可能的改进机会,为本地化代理和客户设计新的自动化和用户体验 (UX) 交互上下文。 确保许可证、部署成本和维护的经济性对我们的业务有意义 关于聊天GPT语言和现实世界用法的几点思考 我发现最引人注目的一件事是ChatGPT如何使复数运算几乎正确,但仍然是错误的。它不会作弊。它真的从它训练的语言中学到了一切。它发现超过某个数量级的操作几乎是正确的结果 (对于较小的数字也是正确的结果),这一事实告诉我们,一个足够规模的语言语料库包含了关于现实世界的统计上显著的知识。但是,它也表明需要专用的形式系统 (例如数学) 来产生有关现实世界的有意义,可靠和准确的信息。 ChatGPT发出了一个清醒的提醒,一个自我参照的、自洽的系统本身不能承载独立存在的世界真相。这与哥德尔的不完全性定理相呼应。作为有意识的生物,我们不能将我们的认知从形式和物质系统中脱离出来,将我们对世界的理解建立在一个强加给我们的现实中,而我们不能仅通过语言来定义。 联系我们 有自己的翻译或本地化项目?是否需要确保它准确、快速且在预算范围内完成?我们将使用像ChatGPT这样的创新生成AI来提供帮助。立即与我们联系,以了解有关Lionbridge的翻译和本地化服务的更多信息。

以上中文文本为机器翻译,存在不同程度偏差和错误,请理解并参考英文原文阅读。

阅读原文