When you are a small localization team, linguistic quality assurance (LQA) is often done on an ad-hoc basis. But, as you start to scale up your localization efforts, you need to put a more formal process in place. You might be working with dozens or hundreds of translators and reviewers, all over the world. How can you keep track of who is doing a good job and who is not?
In this article, we will take a look at what LQA is, how it works, and why it is challenging for large-scale companies. We will also take a look at the multidimensional quality metrics (MQM) framework and how you can automate the process.
What is Linguistic Quality Assurance (LQA)?
LQA is a process of reviewing the source text and its translation to ensure it meets customer standards. Its goals are to check for spelling/grammatical mistakes, correct terminology use, accurate meaning conveyance, style accuracy, cultural adaptation, format correctness, and so on.
Importantly, LQA is a process with more than just one party involved:
buyers should ensure the source text is finalized before sending it to a translation vendor,
LSPs need to screen the source material,
translators must read instructions carefully and, well, translate to the best of their abilities.
Linguistic quality assurance can be broken down into three activities:
source text review,
automated check of machine-detectable errors,
and final check by a native speaker.
For the sake of brevity, we’ll only consider the latter two.
Automated LQA
Automated linguistic quality assurance involves using various software tools to detect typos and errors in the translation. Smartcat, for example, can automatically check for typos and number mismatches, as well as formatting errors. It also has a built-in spellchecker that can be used to quickly identify spelling mistakes.
But, however enticing automated LQA may be, it is not a substitute for manual review. Automated checks can only detect certain types of errors and are prone to false positives. Besides, no automated tool can check for meaning accuracy, stylistical flaws, or cultural appropriateness.
Human LQA
Manual LQA is the process of reviewing translations for errors that cannot be detected by automated tools. It involves a reviewer going through the text and making sure it meets all quality criteria, such as accuracy, style, cultural appropriateness, etc.
However, “quality,” by its very nature, is a very subjective concept. What one reviewer may consider a good translation, another may deem to be of poor quality. This is why it is important to have a well-defined and agreed-upon set of quality criteria, as well as a process for recording errors, as well as a tool for reporting and analyzing them.
Smartcat uses the Multidimensional Quality Metrics (MQM) framework to assess translation quality, so let’s take a closer look at it.
What are multidimensional quality metrics (MQM)?
MQM is a framework for measuring and assessing the quality of translations developed by the World Wide Web Consortium (W3C). In a nutshell, it breaks down quality into several categories, namely terminology, accuracy, linguistic conventions, style, locale conventions, and design and markup (you can read more about them here).
For example, a terminology error could be using “car” instead of “automobile,” an accuracy issue could be mistranslating a phrase, style issues include using too formal or colloquial language (depending on the organization’s style guide), and design and markup errors could be UI element labels that are too long or too short.
Each category has its own weight, which is not standardized and left up to the organization to decide. Each error is also assigned a severity level from Minor (1) to Critical (25), also used for weighting.
The Translation Quality Evaluation (TQE) workflow
Granted, just having some metrics in place is not going to magically improve the quality of your translations. You need to have a process in place for actually using those metrics.
One such process is Translation Quality Evaluation (TQE), which goes side by side with the MQM framework.
In a nutshell, TQE is a workflow that includes the following steps:
1. Preliminary stage, where the metrics are defined and the evaluation criteria are set.
2. Annotation stage, where a human reviewer goes through the translation and marks errors according to the MQM categories.
3. Calculation stage, where a software tool or a spreadsheet compiles a “scorecard” of all the errors and delivers it to the project manager.
The multidimensional quality metrics framework is an industry standard for a number of reasons. First, it provides a common language for everyone involved in the process — the project manager, the reviewers and the translators. It also standardizes the evaluation process, with a clear set of categories and severity levels.
At the same time, MQM is flexible: organizations can decide which categories and severity levels are most important for their particular context. Finally, it mitigates the human factor in quality assurance, with a clear set of rules that reduces the likelihood of arbitrary decisions by reviewers.
How to automate large-scale linguistic quality assurance
While the MQM framework is a great way to ensure quality in large-scale localization projects, there are ways to automate the process and make it more efficient. One such way is to use a tool like Smartcat, which automates the MQM workflow on several levels:
You can create LQA profiles from predefined templates, which include the industry-standard MQM MQM Core and MQM-DQF frameworks:
If needed, you can customize the profiles to better suit your organization’s needs:
The reviewer can add comments with specific MQM categories and severity levels right from the Smartcat interface:
The results are compiled automatically into LQA reports, complete with the overall quality score, a breakdown of errors by category and severity level, and even references to the specific segments where the errors were found:
This way, you can get all the benefits of MQM for large-scale linguistic quality assurance without losing your mind in the process.
当你是一个小型的本地化团队时,语言质量保证(LQA)往往是在临时的基础上进行的。但是,当你开始扩大你的本地化工作规模时,你需要建立一个更正式的程序。你可能要与世界各地的几十个或几百个译员和审稿人合作。你怎样才能跟踪谁做得好,谁做得不好?
在这篇文章中,我们将看看什么是LQA,它是如何工作的,以及为什么它对大规模公司具有挑战性。我们还将看看多维质量指标(MQM)框架,以及你如何能使这个过程自动化。
什么是语言质量保证(LQA)?
LQA是一个审查源文本及其翻译的过程,以确保其符合客户标准。它的目标是检查拼写/语法错误、术语使用的正确性、意义传达的准确性、风格的准确性、文化适应性、格式的正确性等等。
重要的是,LQA是一个不止一方参与的过程:
买方应确保在将源文本发送给翻译供应商之前将其定稿,
LSP需要筛选源材料,
翻译人员必须仔细阅读说明书,并尽最大努力进行翻译。
语言质量保证可分为三项活动:
源文本审阅,
自动检查机器可检测的错误,
和最后检查。
为了简洁起见,我们只考虑后两种。
自动化LQA
自动语言质量保证包括使用各种软件工具来检测翻译中的错别字和错误。例如,Smartcat可以自动检查错别字和数字不匹配,以及格式化错误。它也有一个内置的拼写检查器,可以用来快速识别拼写错误。
但是,无论自动LQA多么诱人,它都不能替代人工审查。自动检查只能发现某些类型的错误,而且容易出现假阳性。此外,没有任何自动化工具可以检查意义的准确性、文体的缺陷或文化的适当性。
人LQA
手工LQA是审查翻译的过程,以发现自动化工具无法检测的错误。它涉及到一个审阅者通过文本,并确保它符合所有的质量标准,如准确性、风格、文化适宜性,等等。
然而,"质量",就其本质而言,是一个非常主观的概念。一个审查员可能认为是好的翻译,另一个人可能认为是质量差的。这就是为什么要有一套明确的和商定的质量标准,以及一个记录错误的程序和一个报告和分析错误的工具。
Smartcat使用多维质量度量(MQM)框架来评估翻译质量,下面我们来详细了解一下。
什么是多维质量度量(MQM)?
MQM是由万维网联盟(W3C)开发的一个衡量和评估翻译质量的框架。简而言之,它将质量分为几个类别,即术语、准确性、语言惯例、风格、地区惯例以及设计和标记(你可以在这里阅读更多关于它们的信息)。
例如,术语错误可能是使用 "汽车 "而不是 "汽车",准确性问题可能是错误地翻译了一个短语,风格问题包括使用过于正式或口语化的语言(取决于组织的风格指南),而设计和标记错误可能是UI元素标签太长或太短。
每个类别都有自己的权重,这不是标准化的,由组织决定。每个错误也被分配一个严重程度,从轻微(1)到严重(25),也用于加权。
翻译质量评估(TQE)工作流程
当然,仅仅拥有一些指标是不会神奇地提高你的翻译质量的。你需要有一个实际使用这些指标的过程。
翻译质量评估(TQE)就是这样一个过程,它与MQM框架并行不悖。
简而言之,TQE是一个包括以下步骤的工作流程:
1.初步阶段,定义指标并设定评估标准。
2.注释阶段,在此阶段,审阅人员将检查翻译并根据MQM类别标记错误。
3. 计算阶段,由软件工具或电子表格编制所有错误的 "记分卡",并交付给项目经理。
多维度质量指标框架是一个行业标准,原因有很多。首先,它为参与这一过程的每个人--项目经理、审查员和译员--提供了一种共同语言。它还使评估过程标准化,有一套明确的类别和严重程度。
同时,MQM是灵活的:组织可以决定哪些类别和严重程度对他们的特定环境最重要。最后,它减轻了质量保证中的人为因素,有一套明确的规则,减少了审查员任意决定的可能性。
如何自动化大规模的语言质量保证
虽然MQM框架是确保大规模本地化项目质量的好方法,但也有办法使这一过程自动化,使其更有效率。其中一个方法是使用像Smartcat这样的工具,它可以在几个层面上实现MQM工作流程的自动化。
你可以从预定义的模板中创建LQA配置文件,这些模板包括行业标准的MQM MQM核心和MQM-DQF框架。
如果需要,您可以自定义配置文件以更好地满足您组织的需要:
审查员可以在Smartcat界面上直接添加带有特定MQM类别和严重程度的评论。
结果被自动编入LQA报告,包括总体质量得分、按类别和严重程度分类的错误,甚至包括发现错误的具体片段的参考。
这样,你就可以获得MQM对大规模语言质量保证的所有好处,而不会在这个过程中失去理智。
以上中文文本为机器翻译,存在不同程度偏差和错误,请理解并参考英文原文阅读。
阅读原文