1 Introduction
The chapter describes major advances in translation technologies (translation memory and
machine translation) and explains how they have influenced our understanding of translation,
particularly the concept of translation quality. The discussion focuses on the notion of
translation as ‘text’, showing that technological changes have created a rift between
Translation Studies theories and a new notion of translation circulating in the industry. The
chapter finally identifies trends in research which seek to develop new knowledge to address
the rift identified.
2 Technologies in Translation Practice
A wide range of electronic tools and systems are available to support the production of
translation (Austermühl, 2014, pp. 18-67). ISO 17100, published by the International
Organisation for Standardization (The British Standards institution, 2018, p. 17), defines
‘translation technology’ as ‘a set of tools used by human translators, revisers, reviewers, and
others to facilitate their work’ and lists the following as examples: a) content management
systems (CMSs); b) authoring systems; c) desktop publishing; d) word processing software;
e) translation management systems (TMSs); f) translation memory (TM) tools and computer
aided translation (CAT); g) quality assurance tools; h) revision tools; i) localization tools; j)
machine translation (MT); k) terminology management systems; l) project management
software; and m) speech-to-text recognition software. Drugan (2013, Chapter 3) provides a
comprehensive summary of the nature of most of these technologies although, inevitably,
some parts (particularly those focusing on MT) need updating due to the innovative nature of
the tools (see below).
This chapter focuses on two technologies used extensively in the standard translation
production processes: translation memory (TM) tools and machine translation (MT).
2.1 Translation Memory (TM)
Translation memory (TM) and machine translation (MT) are often confused, partly because
the acronyms are similar, and also because they have been increasingly used together in a
computer-aided translation (CAT) tool (see below). They were, however, originally
conceived and developed as distinct technologies with unique histories and philosophies
behind them.
We will help you get a quote for your preferred service using our solutions.
Submit your information below for a free, no-obligation quote.
Interlingua Translation provides individually tailored and cost-effective translation services, localisation, post-editing, and transcriptions. Upload your content and ORDER ONLINE
If you would like to know more about our language services or any other request, then CONTACT US and we will be in touch soonest for further deatils.
TM is a key component of a CAT tool. Simply put, TM is a collection of data in the form of a
computer file, which contains sets of source language sentences and their translations in a
target language. Those sets are called ‘segments’. A segment consists of a pair composed of a
source text (ST) sentence and its translation. When translating a new text in a CAT tool with
one or more TMs uploaded, the software compares the sentences in the new ST with the ST
sentences stored in the TM. It detects sentences that are identical or similar in the two, and
shows on the screen the translations of the sentences stored in the TM as ‘translation
suggestions’. The similarity of the ST sentences is measured by the ‘match rate’. A ‘100 per
cent match’ means the two ST sentences (one in the TM and the other in the new text) are
identical. Any match rates smaller than 100 per cent are called ‘fuzzy matches’. The principle
of TM technology is that the more sentences of higher matches the TM contains as against
the new text for translation, the better translation suggestions will be provided, which
facilitate faster translation as the number of corrections the translator has to make is smaller.
As a TM helps translators to translate (but does not produce the translation for translators),
translation produced with a TM (and often with other functions in the CAT tools such as a
terminology tool) is defined as ‘machine-assisted human translation’ (‘MAHT’). For detailed
accounts of TM functions and development, see Bowker (2002) and Bowker and Fisher
(2010).
2.2 Machine Translation (MT)
MT, in contrast, uses a computer system which produces automated translations from one
natural language to another. In the mid twentieth century, MT was originally conceived with
the aim of producing translations without any human involvement, known as ‘fully automatic
high-quality translation’ (‘FAHQT’), but the developers soon realised that this aim was
unachievable. The current practice of MT use is thus defined as ‘human-assisted machine
translation’ (‘HAMT’), as some assistance by humans is necessary to achieve high quality
translation.
The methods used in MT systems have changed over time. Broadly speaking, there are two
approaches: rule-based and data-driven.
Rule-based machine translation (RBMT) is a system built by programming grammatical and
lexical rules of both languages. Because of limitations in the quality of the target language
output of rule-based systems, since the 1990s, MT development has gradually moved toward
data-driven approaches. Data-driven MT systems are built by letting the system learn patterns
of translations from a large number of parallel corpora consisting of ST sentences and
translations of them made by humans. The system produces translations by calculating the
statistically most probable translation. The two main methods used are statistical machine
translation (SMT) and the more recently developed neural machine translation (NMT). Most
major MT systems (including free online ones like Google Translate and Microsoft
Translator) now use NMT. SMT and NMT both use data-driven approaches, but NMT uses
artificial neural networks as an analysis method and is said to be capable of producing more
fluent translations, although it has its own limitations such as a tendency to produce
semantically inaccurate passages or non-words.
In this chapter, ‘MT’ refers to data-driven MT systems because these are the mainstream
systems. For detailed accounts of the history of different MT systems, see Kenny (2018). For
the basic working principles of NMT, see Forcada (2017).
2.3 How TMs and MTs are Used in Practice
Since the mid-1990s, many CAT-toolproducts have become more affordable for translators,
and practicing translators are often expected to use TMs in the CAT-tool environment to
speed up translation processes. The effectiveness of TM use for translation productivity is
affected by several factors, such as the number of suggestions the TM can offer and their
match rates. The effectiveness is also influenced by the nature of the ST: translations of texts
that follow a standard format (such as users’ manuals, legal texts and IT documents) are more
With translation featuring prominently on the agenda of today’s multicultural and multilingual society, the need for translators will only increase. We at Interlingua Translations, are proudly part of this inevitable trend.
Interlingua Translation Limited is devoted to the weight of words and the power they have to better our world. So, if you are looking for a steadfast, responsive and viable freelance translator, then this platform is right up your alley.
Bumping into an appropriate Translation service provider in the net can be a bit like a needle in a haystack if you are not a Linguist or have very limited knowledge of the translation industry. At Interlingua Translation Limited.....
likely to achieve high productivity with TM. TM is also effective when only some parts of
the ST need translating (such as an updated version of a user’s manual) as the CAT tool can
pick up the sentences which need updated translations while it reuses previous translations
for 100 per cent matches (source text sentences which have not been changed from the
previous version).
Due to their typically imperfect quality, MT outputs often undergo a process called ‘machine
translation post-editing’ (‘MTPE’), in which human operators (often called ‘post-editors’ or
‘linguists’) correct mistakes in the MT outputs. Due to the low quality of MT in the early
development period, MTPE was limited to producing gist translations for corporations’
internal uses (Garcia, 2011, p. 218). However, since the quality of data-driven MT systems’
output has improved, MTPE has become widely used in the commercial translation market.
For a description of a typical MTPE process in the industry, see Zaretskaya (2017).
Although TM and MT were originally conceived and developed for different purposes, as
explained above, the boundary between the two technologies has become blurred because of
a function called ‘MT assisted TM’ (Garcia, 2010), which many CAT tools incorporate. This
function offers MT outputs as translation suggestions when TMs cannot offer good enough
suggestions. For example, a user can calibrate the CAT tool so that if matches of higher than
70 per cent are not offered by the TMs, the tool will retrieve MT outputs from a specified
online MT system via a plugin, a piece of software code which allows the CAT tool to access
and use an external MT system. When using this function, the translator’s translation process
is assisted by TMs and MTsat the same time. The use of two kinds of leverages (TM
matches and MT outputs) is believed to improve translation efficiency.
3 The Influence of Technologies on the Understanding of Translation Quality
3.1 The Sentential Approach to Translation
TM and MT share as a fundamental working principle the use of the sentence as the unit of
translation. A TM holds its data in the form of a parallel corpus, in which each segment consists of a
sentence from the ST and its translation in the target language. ‘A sentence’ here means a
string of words which ends with a full stop. A segment can consist of other text units such as
a phrase or even a word (e.g., in a chapter heading in a text), but for ease of argument, we use
the word ‘sentence’ here.
A CAT tool retrieves translation suggestions from TMs for each sentence. A CAT tool’s
interface is designed so that the translator translates sentence by sentence: the translator
evaluates the suggestions for a current sentence (the sentence in which the cursor is placed
and is highlighted on the screen), chooses the suggestion he or she considers the most useful,
and amends the sentence as necessary. If the TMs do not offer any usable suggestion, the
translator rejects the suggestions and translates the sentence from scratch. Once the translator
‘confirms’ (accepts) the translation, the CAT tool highlights the next segment, showing
suggestions. This procedure limits the translator’s cognitive processing of the text to the
sentence level, making cross-sentential operation difficult.
With MT, too, the unit of translation is a sentence. The MT algorithm recognises one source
language (SL) sentence as a unit of translation and produces a translation of that sentence.
Text production is thus achieved by accruing translated sentences in a linear mode.
Assessment of the quality of MT outputs is also carried out at a sentence level. There are two
types of MT assessment: automatic and human. In automatic assessment, the quality of MT
outputs is typically measured using human translation as a benchmark. The implicit
assumption in automatic assessment is that quality can be assessed by the similarity to the
human reference translation at the word or sentence level. The closer the MT output is to the
reference translation in terms of the number of identical words included in the sentence, the
higher a score it achieves. In human assessment, human assessors assess each sentence of
machine translation output by using their judgement of quality with regard to different
criteria (such as accuracy and fluency). This assessment is also carried out at the sentence
level, i.e. without a context. For a summary of these assessment methods, see Doherty (2017).
In the context of MTPE, efficiency is related to the amount of effort spent to correct errors in
raw MT outputs, so knowing the number and types of errors in MT outputs is important. A
body of MT research has identified a number of different error classifications and techniques
of error analysis (Popović, 2018), again carried out at a sentence level (or at the level of a
segment within a sentence).
3.2 Catford – the Sentence-Bound Linguistic Approach to Translation
This sentence-focused principle of translation assessment reminds us of Catford’s early
linguistic approach to translation (1965). Because translation, according to Catford (1965, p.
1) is ‘an operation performed on languages’, a theory of translation needs to draw on a
linguistic theory, and Catford selects the General Linguistic Theory presented by Halliday
(1961). The highest rank of this grammar is (as in most grammars) the sentence, and Catford
considers that it is at this grammatical level that translation equivalence can most often be
established:
SL [source language] and TL [target language] texts or items are translation equivalents
when they are interchangeable in a given situation. This is why translation equivalence
can nearly always be established at sentence-rank—the sentence is the grammatical unit
most directly related to speech-function within a situation. (Catford, 1965, p. 49)
(italics in original). That said, his model allowed for shifts, ‘departures from formal correspondence’ between levels and between categories (Catford 1965, p. 73). For example, a sentence can be
translated into a phrase or a word, and vice-versa, which constitutes a level shift. A category
shift occurs when structures, classes and units do not correspond formally between a ST and a
TT. For example, a clause structure shift takes place when the English clause ‘John loves
Mary’, which has the structure Subject – Predicator – Complement, is translated into Gaelic
as ‘Tha gradh aig Iain air Mairi’, which has the structure Predicator – Subject – Complement
– Adjunct (Catford 1965, pp. 76-77).
3.3 The Textual Approach to Translation
The notion of translation equivalence at the sentence level and the observation of its
manifestation in the target text has limitations. For example, it does not explain what the
translator really does when translating (Fawcett, 1997, p. 56). Since the late twentieth century,
Translation Studies scholars have drawn on other branches of linguistics which had
developed by then to explain the real-world experiences of producers, readers and translators
of texts. These include, among others, text linguistics. ‘Text’ is understood in different ways
by different schools of linguistics, often being interchangeably used with ‘discourse’. The
concepts were famously imported into translation by Hatim and Mason (1990) from De
Beaugrande and Dressler (1981).
De Beaugrande and Dressler (1981, p. 3) maintain that a text is a communicative occurrence
which meets seven standards of textuality: cohesion, coherence, intentionality, acceptability,
informativity, situationality, and intertextuality. These seven standards are all necessary to
make the text ‘communicative’, but the standards that are particularly relevant in the current
discussion are cohesion and coherence. Cohesion in text is produced by certain linguistic
devices, which are language specific. For instance, English has five such devices: reference,
substitution, ellipsis, conjunction and lexical cohesion (Halliday and Hasan, 1976). These
devices link different parts of text together, which enables the reader to recognise one part of
a text as connected to another part of the text. However, textuality involves another important
concept, coherence, because a text must hang together, ‘both linguistically and conceptually’
(Hatim and Mason, 1990, p. 192). For it to hang together conceptually, it needs coherence as
well as cohesion. Coherence is not an observable textural feature, but, instead, is produced by
assumptions held by the reader. To recognise a set of linguistic devices as a vehicle for
creating coherence, the reader needs to have world knowledge along with knowledge of
coherence relations such as cause-effect, problem-solution, and temporal sequence. Both
cohesion and coherence are standards of text (Hatim and Mason, 1990, p. 195).
Application of these concepts are not straightforward in translation, though. This is because
while the knowledge of the sequence of coherence relations is (most likely) shared by users
of the source and target languages, cohesive devices in those languages are language specific.
The translator’s job is to negotiate between the two systems and produce translation which
satisfies both systems and achieves equivalence at text level (Baker, 2018/1992, pp. 134-234).
As Shreve puts it, paying attention to textual properties can lead to a ‘quantum leap ... in the
“textual quality” of the translation’ (Shreve, 2017, p.175); but text production through
negotiating between two systems (i.e. translation) requires considerable linguistic knowledge
and skills and major cognitive effort. This is because the unit of translation is considered to
be a sentence (Huang and Wu, 2009) and, as a body of cognitive research indicates, the way
that translation proceeds (i.e., sequential and step-by-step) brings the translator’s focus,
necessarily, down to the sentential level (Shreve, 2017, p.175). This leads us to deduce that
producing a cohesive and coherent text on the text level in translation requires additional
effort and skills (such as cross-sentential revision) by the translator compared to mono-
lingual text production.
4 The Influence of Technologies on Translation Practice
4.1 The Influence of TM on Text Production
Considering the cognitive processes involved in translational text production, as well as the
mechanism of the TM function in a CAT tool, it is easy to understand that TM-assisted
translation exacerbates the difficulty of producing translation with appropriate textuality. The
translator is required to minimise the adverse effects caused by the sentential restrictions as
well as the influence from the CAT tool interface and the nature of TM. Under these
circumstances, the translator must pay sufficient attention to translation suggestions from the
TM and apply necessary edits so that the final translation achieves maximum textuality.
Bowker (2006) offers detailed examples of the difficulties involved in this. One obvious
example is related to the sequence of coherence relations. A sentence in a text stands in
certain logical relations to other parts of the text, but these relationships obtain only in that
particular text-context configuration. When a sentence is stored in a TM, the context is
stripped away. And when the sentence is suggested as a recyclable segment in a new
translation, the suggestion is not only made without regard to context, but may conflict with
the new context in the target text (TT). Therefore, to make the suggested sentence work in the
new translation, the translator needs to ‘work outside the artificial boundaries of sentences, so
the sentence-by-sentence approach imposed by TMs may not be conducive to effective
translation of the text's message as a whole’ (Bowker, 2006, p. 180). Another example
concerns polysemy. Bowker uses the example of the French translation of the English
expression ‘empty the pipe’, which will have different meanings in a text about plumbing
(pipe=tuyau) and in a text about smoking (pipe=pipe) (Bowker, 2006, p. 179). Since a
suggestion from the TM database is presented out of context, the translator will need to check
the context, and if necessary, consult relevant reference materials.
These examples indicate that, although the use of TMs was originally introduced to improve
the productivity of translation processes, the resultant translations require a high level of
effort by the translator to ensure that the suggested sentences are adjusted so that they sit well
in the whole text. In addition, the effort to accomplish this task is influenced by some TM-
related factors, typically, ‘terminological train wreck’ and ‘sentence salad’ (Bowker, 2006,
p.181). ‘Terminological train wreck’ occurs because translators tend to use inconsistent or
inappropriate terminology for the context of a new translation when the TM contains
terminology from different translations made for different clients at different times.
Terminology can evolve quickly; thus terms stored in a TM may become out of date quickly.
‘Sentence salad’ is a similar phenomenon, but at the sentence level. If a TM offers translation
suggestions from translations produced by different translators in different styles about
different topics, the new translation based on suggestions from the TM may present a
‘stylistic hodgepodge’.
Dragsted’s (2006) process study investigated how professional and student translators deal
with these challenges posed by TM. The study showed that, when translating without a TM,
professional translators tended not to recognise the sentence as a translation unit (whereas
student translators did tend to translate sentence-by-sentence), but when translating with a
TM, their focus was more sentential: they spent more time translating each sentence,
reducing the length of time spent on cross-sentential revisions at the end of the whole
translation (Dragsted 2006, pp. 449-453). The professionals also made fewer cross-sentential
shifts from the source text to the target text, such as combining or splitting up sentences,
when using a TM (pp. 453-459). Most professionals recognised this cognitive restriction
caused by the TM mechanism as a disadvantage, while student translators tended to regard it
as an advantage, saying that they could concentrate on translating one sentence at a time.
These findings suggest that student translators do not recognise the restrictions caused by a
TM because their translation competence is not sufficiently developed to deal with the text-
level translation problems when translating without a TM (p. 457). In contrast, professionals
are aware of these restrictions, but in the study, professionals said solving text-level
translation problems at the final cross-sentential revision stage was not always possible in
professional situations because it requires time and effort (p. 458).
These findings provide a convincing explanation for another TM-related phenomenon called
‘blind faith’ (Bowker, 2005). ‘Blind faith’ refers to the tendency of translators to accept
suggestions from a TM without checking the appropriateness of the segment in the new
translation sufficiently due to the pressure to increase productivity. The pressures for higher
productivity in the professional environment is highlighted in translation workplace studies
(e.g. Le Blanc, 2017), which show that translators are increasingly deprived of time to spend
on translation to meet management’s expectations that technologies like TM should reduce
the time spent on translation.
Our expert linguists can help you effectively communicate with your clients while meeting insurance-related regulationsin translating insurance-related reports and documentation materials. FIND OUT MORE.
Our expert team of linguists understand even the most complicated educational texts and terminology, thus helping you to provide accurate study materials for your non-English speaking students. FIND OUT MORE.
Interlingua translation has the skills and experience to help you get what you need from your government translation services. So, no matter how niche your request is, our team will be able to help. FIND OUT MORE.
Maintaining accuracy and nuance in your translated marketing materials can make or break their impact internationally. INTERLINGUA TRANSLATION is your answer. Unlock your interest and FIND OUT MORE.
4.2 The Influence of MT on Text Production
The rift between the concepts of translation quality at the text level and at the sentence level
is evident in text production involving MT, too.
Lumeras and Way (2017), drawing on Kay (2014), illustrate the rift using the dichotomous
concepts of ‘syntactic translation’ and ‘pragmatic translation’. Lumeras and Way (2017, p.
30) provide the following example.
(i) ‘Est-ce que c’est ta cousine?’ ‘Non, je n’ai pas de cousine.’
The cousin who is talked about in this conversation is female, which is clearly marked by a
female form of the word ‘cousine’ in French. However, as the English language does not
have lexical items to distinguish male and female cousins, an MT engine is most likely to
render the French sentences like (ii).
(ii) “Is that your cousin?” “No, I don’t have a cousin.”
On the other hand, if the conversation is to be translated by a human translator, the translator
will infer the gender of the cousin from the other parts of the text, or from their world
knowledge about the text, and may produce a text like (iii).
(iii) “Is that girl your cousin?” “No, she’s not my cousin.”
A sentence like (ii) is called a ‘syntactic translation’ and a sentence like (iii) is called a
‘pragmatic translation’. On the sentence level, (ii) is correct, but is not the optimal translation
on the text level.
MT translates a text which consists of more than one sentence through a production model
that Lumeras and Way (2017, p. 30), following Kay (2014), call the Syntactic Model of
Translation. The model bases itself on the idea that ‘a long translation is a sequence of
short(er) translations, we memorize short translations (lexical items), and these short
translations can be reordered’ (p. 30). This model is obviously incongruent with the notion of
text we saw above.
The adherence to sentence-level text production is evident in MTPE training, too. A piece of
training material of one of the major localisation companies (SDL, 2017, p. 32) advises:
1. Read the source segment first then the MT output.
2. Determine the usable elements (single words and phrases) and make them the basis for
your translation.
3. Build from the MT output and use every part of the MT output that can speed up your
work.
The guidelines then suggest that all grammatical and terminological errors be corrected
whereas corrections of stylistic errors are optional. After editing MT outputs to the end of the
text in this way, the post-editor is instructed to run the automatic quality check function
available in the CAT tool, which detects spelling, grammar and terminological errors. There
are, however, no instructions in these guidelines to check text-level errors (cross-sentential
checks). This indicates that quality assessment on the text level is not expected in MTPE. In
the same vein, the post-editing guidelines published by the Translation Automation User
Society (TAUS) (Massardo et al., 2016, p. 17) encourage post-editors, while verifying edits,
to ‘[e]nsure that no information has been accidentally added or omitted [to the source
segment]’. The reason behind these instructions is easy to deduce: trying to assess the quality at text
level, which may necessitate additional cross-sentential translation strategies, will make the
exercise complex and time-consuming, which defeats the purpose of MTPE, i.e. fast
turnaround. This sentence-focused process in MTPE tends to cause text-level problems just like
translation using TMs does. Čulo, Gutermuth, Hansen-Schirra and Nitzke, (2014, p. 208)
report a case of English-German post-editing. Killer nurse receives four life sentences. Hospital nurse C.N. was imprisoned for life today for the killing of four of his patients. (Source Text)
Killer-Krankenschwester zu viermal lebenslanger Haft verurteilt. Der Krankenpfleger
C.N. wurde heute auf Lebenszeit eingesperrt für die Tötung von vier seiner Patienten.
(Post-edited text). ‘Killer woman-nurse to four times life-long imprisonment sentenced. The man-nurse C.N. was today for lifetime imprisoned for the killing of four of his patients.’ (Back
Translation). In the first sentence, which is a newspaper headline, the word ‘nurse’ in the English source text was machine translated as ‘Krankenschwester’ (female nurse), which the post-editor
accepted. In the following sentence, however, the post-editor edited the same word to
‘Krankenpfleger’ (male nurse) as the gender of the nurse was clarified by the pronoun ‘seiner’
(=his) in the same sentence. The inconsistency across the two sentences is not fixed in this
example, leaving the text incoherent at the text level. Čulo et al. (2014, p. 212) conclude that
the post-editing task required the post-editors to use the translation strategy of ‘explicitation’.
4.3 Justification from the Industry
We have seen that the translator’s ability to produce high-quality translation text tends to be
affected by the use of both TM and MT. These tools turn the process of translation into a
mechanical task of producing and assessing translation sentence by sentence, which can be
considered to be side effects of the technologies. When the phenomena are considered in
sociological frameworks, the side effects of TM and MT can be understood as a burden
imposed on translators by tool makers and employers/commissioners of translators who
prioritise the operationality of technology over and above the importance of the concept of
‘text’. As a result, the translator’s pursuit of quality translation as text is subjected to ‘the
industrialization and globalization forces that demand higher productivity and speed’
(Jiménez-Crespo, 2017, p. 161). Despite the importance of the concept of text as the basis of
many new translation research and training projects, the concept is undermined in practice by
technologies such as TM and MT (Jiménez-Crespo, 2017, pp. 158-9).
However, this does not stop the industry from using these technologies in translation. A
common ground for a defence of this practice is the notion of ‘good enough translation’ or
‘fit-for-purpose translation’ (Bowker, 2019). When time and budget resources are limited,
one of the strategies used in the industry is to target resources according to the purposes of
translation. If the text is highly ‘perishable’, i.e. to be used only for a short time, perfect
translation may not be necessary. Texts such as forum posts on social media or product
support texts (which are updated frequently) are good examples. Another criterion for
resource allocation is the risk involved in the translation (Nitzke, Hansen-schirra, and
Canfora, 2019). Outward communication from a company to their customers is important for
the business and the risk of damaging the business is high with translation of texts of, for
example, advertisement. In contrast, texts used for internal information carry a lower risk.
Users of translation may decide to have perishable and lower-risk texts translated at a lower
quality, i.e. ‘good enough’ translation.
The concept of ‘good enough’ translation manifests itself clearly in the two-tier quality
classification of MTPE, which is commonly used in the industry. The classification consists
of two distinct post-editing processes: ‘full post-editing’ and ‘light post-editing’. ISO 18587
(The British Standards Institution, 2017), an international industry standard which sets
standards of MTPE services, defines ‘full-editing’ as editing required ‘to obtain a product
comparable to a product obtained by human translation’ and ‘light post-editing’ as required to
‘obtain merely comparable text without any attempt to produce a product comparable to a
product obtained by human translation’. The concept of light post-editing embodies the
industry’s belief that a substandard translation has its own market value, depending on the
clients’ needs and requirements, budgets or time constraints.
The discrepancy in the fundamental attitudes to quality between academic theories and
professional beliefs is evident in Drugan’s (2013) extensive ethnographic study of translation
quality in the industry. Drugan points out that no professionals or practitioners who
participated in the study mentioned a single translation studies model to explain their practice
of quality assessment: academic theories are ignored in professional environements (p. 41).
At the same time, Drugan points out that much of the industry debate on ‘fit-for-purpose’
translation, for instance, is clearly linked to ideas from Skopos theory, even if this is rarely
acknowledged, or perhaps even realized (p.47).
Skopos theory (Reiss and Vermeer, 2013/1984) postulates that texual features of translation
are governed by the purpose (=skopos) of translation. The purpose of translation is decided
by the situation, which includes what the recepient of the translation expects from it (p. 89).
Consequently, the message in the translation should be ‘“sufficiently” coherent with the
situation in which it is received’, which is more important than the massage being ‘coherent
“in itself”’ (p. 98). ‘“Understanding” means to relate something to one’s own situation and
the background knowledge it implies’ (p. 98).
Although the practice of MTPE was not
common when Skopos theory was developed, this reasoning, particularly the understanding
of ‘coherence’, agrees with the practice of MTPE. For example, a mechanical engineer will
find a light post-edited technical document about a new machine sufficient for their purpose
as they can use their own situation (their own world knowledge about the subject) in
interpreting the text correctly. Light post-editing of
MT will be adequate for the purpose.
So, although the notion of translation commonly observed in the technological industry
environment seems to conflict with the academic concept of translation quality as determined
by the quality of the translated text in the context of its source text, the notion used in the
technological environment shares much common ground with one of the most authoritative
translation theories, i.e. Skopos theory..
4.4 Future Outlook for Technologies and Translation Studies
We have seen above that technologies such as TM and MT are not capable of addressing the
notion of text sufficiently due to their sentence-focused machnisms. One may hasitily deduce,
then, that, as technologies develop further, the gap between the understanding of translation
in Translation Studies and in the industry will widen. In reality, however, this seems not to be
common when Skopos theory was developed, thisreasoning, particularly the understanding
notion of text in translation technology development.
With regard to TM, most CAT tools incorporate a function called ‘preview’. A separate
preview pane on the computer screen shows the target text in the form it will appear on the
final printed page. With this function on, the translator can check how the translation looks
on the page, i.e. on the text level in the intended context. This function facilitates the
assessment of the level of cohesion and coherence in the translation.
In MT research, an increasing number of investigations are carried out with a view to
improving the text-level quality of MT outputs. This includes the development of translation
and quality assessment models which can refer to discourse information from outside the
current sentences. At the time of writing, an increasing number of research projects are
exploring methods to materialise this (e.g. Bawden, Sennrich, Birch and Haddow, 2018; Li,
Nakazawa and Tsuruoka, 2019; Wang, 2019). This shows that language technology research
in the computing and engineering disciplines are actively importing the notion of quality
from Translation Studies.
This kind of text-level MT research is still in its infancy and how much of this goal can be
achieved remains to be seen. Also, how this new stage of MT development influences the
perception held by users of MT, as well as translators, will be an important focus for
observation. It may be welcomed by the translation community as a sign of diminishing
polarisation and opposition betweeen humans and machines, a common trait seen thus far in
the translation community (Sakamoto, 2019a).
On the other hand, this may exercerbate the
feeling of threat experienced by translators if it is understood as a sign that MTs are catching
up with humans and will eventually make human translators obsolete. The relation between
translation and technology is fluid, and if or how our understanding of translation will be
changed by technologies in the future remains to be seen.
5 Future Directions of Research
O’Hagan (2013, p. 508) points out that there is a disconnection between the theory and
practice of translation, claiming that the changes occuring in technologies and their effects on
translation practice, and the studies in the applied branches of translation which investigate
these phenomena, are not influencing theorising or modelling in the pure branch of
translation studies on Holme’s map of Translation Studies (Holmes, 1972). The examples
offered in this chapter are relevant. As we have seen, methods and evidence used in
technology-related research in the applied branch have been restricted to the sentence-level
due to technological limitations. Text-level engagement is, however, increasingly becoming
possible thanks to technological advancement, which promises increased interaction between
the pure and applied branches.
Diverse approaches in studies of translation, including both cognitive and sociological studies,
are needed in the pursuit of such interaction, and to make the outcomes useful for the real
world. The influence of technology on translation, both positive and negative, is enormous
and consequently affects all parties involved in it (i.e., translators, post-editors, business
owners, clients of translation services, etc.). What follows shows some examples of such
studies and future directions of research.
concurrent and retrospective verbal reports (Göpferich and
Jääskeläinen, 2009). The processes of translation with tools such as TM and MT have been
an important target of such studies. Christensen and Schjoldager (2010) and Christensen
(2011) offer a good review of studies of TMs. More recently, in tandem with the increasing
prevalence of MT and its integration in CAT tools, the number of process studies with MT
has increased, particularly concerning MTPE processe
5.1 Cognitive Studies of TranslaTion Technologies
Process studies have been popular in Translation Studies since the 1980s. Methods such as
think-aloud protocols (TAPs), keystroke logging, screen recording and eye-tracking are
typically used, as well as concurrent and retrospective verbal reports (Göpferich and
Jääskeläinen, 2009). The processes of translation with tools such as TM and MT have been
an important target of such studies. Christensen and Schjoldager (2010) and Christensen
(2011) offer a good review of studies of TMs. More recently, in tandem with the increasing
prevalence of MT and its integration in CAT tools, the number of process studies with MT
has increased, particularly concerning MTPE processes. For a review of such studies, see
Koponen (2016). For collections of studies on MTPE, see O’Brien (2014), O’Brien and
Simard (2014) and Vieira, Alonso and Bywood (2019).
One of the main questions in MTPE research is whether MTPE is worth adopting (Garcia,
2011; Koponen, 2016). To answer this question, studies have examined the quality of
translations produced by MTPE, their difference from human translation and their efficiency.
But such investigations tend to measure quality in terms of the number of errors, and
inevitably face the fundamental questions: ‘What is quality in translation? How can it be
measured?’.
An innovative approach to this question is to investigate ‘post-editese’, a concept derived
from ‘translationese’. Daems, De Clercq and Macken (2017) investigated whether post-edited
products carried more typical MTPE features and concluded that sufficiently post-edited MT
outputs do not carry post-editese features which machines can detect. In contrast, Toral
(2019) investigated lexical variety in MT outputs as well as post-edited texts and argues that
post-editese is observable in post-edited texts. These studies are good examples of how a
traditional Translation Studies concept (in this case, translation universals) and new practices
involving technologies can be linked to explore the question of quality.
Another noticable development in MT research is the emphasis on ‘prediction’ of MT quality,
dubbed the ‘predictive turn’ in Translation Studies (Schaeffer, Nitzke and Hansen-Schirra,
2019). Accurate prediction of MT output quality will allow us to filter out suitable MT
outputs for an effient MTPE process. Furthermore, researchers are interested in estimating the
cognitive effort required in post-editing by certain features of texts. Outcomes of such studies
will have applied value in professional workflow design and management, including time
planning and economic modelling of MTPE services.
5.2 Sociological Studies on Translation Technologies
The other side of the coin of translation technology studies is the sociology of translation. To
understand the influence of technologies on translation practice and on the people involved
with it, researchers have imported different sociological frameworks into Translation Studies.
For example, in studying the way users’ agency resists tool development, Olohan (2011)
adopts Pickering’s notion of a ‘Mangle of Practice’ (Pickering, 1993). In an attempt to
answer why many translators resitst post-editing work, Sakamoto (2019b) draws on Bourdieu’s field theory (Bourdieu, 1984/1974). In examining how stakeholders’ legal rights
and power relations are affected by the way translation resources (MT’s machine learning
data) are used, Moorkens and Lewis (2019) use Hess and Ostrom’s institutional analysis and
development (IAD) framework (Hess and Ostrom 2007). These and a growing number of
related works can be grouped together under an umbrella paradigm of ‘Technology and
Science Studies (TST) inspired Translation Studies research’ (Kenny, 2017; Olohan, 2017;
Sakamoto, Evans, and Torres-Hostench, 2018). This research paradigm examines relations
and interactions between technology and translation from critical angles using historical,
economic, sociological and antholopological methodologies.
It is important to remember that the two approaches (the cognitive and the sociological) do
not, or should not, position themselves seperately from each other if we are to achieve a
holistic understanding of translation in the technologised society and benefit from it in the
real world. One good example of such holistic enquiry is workplace research. Because it uses
controlled, experimental methods, cognitive research has limited ecological validity. To
overcome this limitation, some researchers are going out of the labs to examine the
interactions between humans and technologies in workplaces. This approach may be
described as ‘socio-cognitive’ (Risku, Rogl and Milosevic, 2017), ‘socio-technical’
(Ehrensberger-Dow and Massey, 2017) or ‘cognitive-ergonomic’ (Lavault-Olléon, 2011;
Teixeira and O’Brien, 2017), but they all investigate how technologies influence and shape
practice in the real world, and vice versa. These interactions between different branches of
Translation Sudies have the potential to lead us to yet further enhanced understanding of
translation.
This paradigm covers a wide range of topics in translation. These include technology-induced
power relations between different stakeholders (e.g. Garcia, 2007), economics of translation
(e.g. Moorkens, 2017; Vieira, 2018), ethical use of data resources (e.g. Drugan and Babych,
2010; Kenny, 2011), and translators’ perceptions of technology use (e.g. Guerberof Arenas,
2013; LeBlanc, 2017). There are also fields of
We will help you get a quote for your preferred service using our solutions.
Submit your information below for a free, no-obligation quote.
Interlingua Translation provides individually tailored and cost-effective translation services, localisation, post-editing, and transcriptions. Upload your content and ORDER ONLINE
If you would like to know more about our language services or any other request, then CONTACT US and we will be in touch soonest for further deatils.
Interlingua Translation Limited
12 The Weavers, Swindon, SN3 1SE, United Kingdom
Copyright © 2025 Interlingua Translation Limited - All Rights Reserved.
Powered by GoDaddy