Terminology in The Age of Artificial Intelligence: François Massion
Terminology in The Age of Artificial Intelligence: François Massion
Terminology in The Age of Artificial Intelligence: François Massion
87–108
© 2021 François Massion - DOI https://doi.org/10.3726/JTS012021.6
françois massion
D.O.G. Dokumentation ohne Grenzen GmbH
francois.massion@dog-gmbh.de
Abstract
Until a few years ago, artificial intelligence (AI) played no role in practical terminology
work. Meanwhile, intelligent terminology management systems have become part of the
tool landscape. They model knowledge by building concept networks with the help of rela-
tions. These intelligent terminologies can help improve the performance and quality of AI-
based applications and support translators and interpreters in their daily work. Intelligent
terminology repositories offer challenging new opportunities for translators, interpreters
and knowledge workers. This article gives some background information and explains the
functioning of intelligent terminologies.
Keywords
artificial intelligence, machine translation, terminology, semantic frame, machine cognition
1. Introduction
Journal of Translation Studies vol. 01/2021 - This work is licensed under a Creative Commons
CC-BY 4.0 license. To view a copy of this license, visit https://creativecommons.org/licenses/by/4.0/
88 François Massion
common understanding of the term Brexit (or associated terms like backstop
or hard border, etc.). This is representative of many situations experienced
by language professionals.
As terms play a key role in translation and interpretation and in
communication, many organizations, companies or individuals have built
multilingual terminologies that they store in different formats, from simple
Excel file to complex relational databases. With the rise of AI, some of the
established ideas on terminology are now under review.
So far, the ideas of Wüster’s critics have had little impact on the
terminology repositories used by professional translators or interpreters.
Most terminology repositories in the industry are concept-based. This
means that for each language all terms designing a concept are gathered
and recorded together with attributes and definitions. In many cases these
repositories contain usage recommendations, i.e. they display allowed and
prohibited terms. The reason for this approach can easily be understood
from a practical point of view knowing e.g. that large industrial companies
or organizations frequently record numerous terms for the same part (car
phone, car telephone, cell phone, cellular phone, mobile phone, etc.) and
want to standardize their corporate language.
3. AI disruption
But all these developments took place in a parallel world, far away from
the daily routine of interpreters and translators. As it is too often the case,
sciences thrive in silos and sometimes research is done independently and
in parallel on similar subjects.
Now AI has reached the radar screen of translators, interpreters or
language professionals. A turning point has been the widespread use of
Deep Learning and Convolutional Networks since 2012 (Le Cun 2019) that
boosted the performance of machine translation. Since then, most language
professionals and scholars have been struggling to assess what effects AI will
have on their profession. Artificial Intelligence has had a disruptive effect
in nearly all fields where it is applied. And terminology is no exception.
Until now the major purpose of terminology work has been to model
and construct terminology repositories for use by humans. Whatever the
prevailing theory was, its purpose has been to serve humans. The task of
modelling knowledge for machines had been tackled separately by other
disciplines like computer science with research areas such as logic (building
Terminology in the Age of AI 91
up on the early work of Gottlob Frege and Bertrand Russell1) and ontologies
(Willard Quine 2).
AI led to the introduction of new types of terminology repositories into
the traditional tool landscape. Intelligent terminologies a.k.a. ontotermino-
logies, knowledge- rich terminologies, relational terminologies, termonto-
logies are right now available for translators and interpreters. Intelligent
terminologies belong to the family of augmented translation technologies
that are inspired or driven by artificial intelligence. They model knowledge
by creating multilingual conceptual networks using relationships.
Like other technologies and concepts such as AI or machine translation,
the idea is not new. Ontoterminology is a term coined by Roche in 2007
and further developed (Roche 2007; Roche et al. 2011) and termontology
or termontography can be traced back to Temmerman and Kerremans in
2003 (Temmerman and Kerremans 2003; Kerremans and Temmerman
2004). Besides, the idea of connecting concepts has been formulated long
ago. It can be found in Ross Quillian’s Ph.D. thesis in 1966 that introduced
the idea of semantic networks (Quillian 1968). The difference now is that
these ideas have become part of practical applications that are available
to be used today on a wide scale. A look at WIPO Pearl3, BABELNET4 or
tools like LOOKUP5 can confirm this.
1. First the term used or the referent. It is the entry point into the triangle.
A term is supposed to perfectly link the abstract concept and the ob-
ject of the concept. But language is ambiguous and imprecise. There
is polysemy and individuals use words that do not always cover the
semantic features of the abstract definition. They often use a more
general term. The meaning of the word comes from its association
with other words or from extralinguistic aspects. Fillmore describes
e.g. how in certain circumstances terms like magistrate and judge
are used as synonyms. The synonymy is given by the semantic frame
(Fillmore and Baker 2001).
2. Next, the object of the concept itself, that is an individual instance
of the concept. This object can be real and abstract and may have
deviating individual properties due to the context, the situation or the
individual interpretation of a person, etc.
Terminology in the Age of AI 93
3. Third the definition itself. Definitions are not cast in stone. They are
influenced by several factors and can vary substantially depending
on the source of the definition. First of all, reality is not perceived the
same way by everyone: culture, language and individual experien-
ce play an important role. The purpose of terminology work is also
important. While in many cases, terminologists who create entries
de-contextualize concepts and formulate definitions that are intended
to be valid for as many users as possible, in other cases they pursue
specific legitimate goals when they select and define concepts. As
an example, you can look at a bike as a means of transportation, as
a tool to improve your physical condition or as an item in the ledger
of a company that sells products to customers. Depending on the in-
tention, the elements of the definition and the equivalents in other
languages will vary.
Thus, the semiotic triangle does not offer the flexibility to describe “soft”
and variable features of a concept, for instance the communication si-
tuation, the context, the objectives, the experience, the culture, etc. The
semiotic triangle itself is static and represents a frozen in vitro definition
of a concept.
This is the reason for which terminology entries do not always reflect
the actual situation in which a term is used. The connection between the
definition of the concept, the term and the object or idea the translator is
working on fails. Translators or interpreters do regularly experience this
and often need to consult several terminological sources before finding a
translation or definition that is acceptable.
Most terminology databases in use today do not provide a mechanism
to respond to different usage situations. They leave translators and inter-
preters alone with their translation issue. For example, what is the correct
translation of the word container, knowing only the definition “an object for
holding or transporting something”. Unless we see the container or have a
detailed description of the situation, we have no chance of knowing exactly
what it is. It can be a box to transport parcels or used to ship goods overseas
or a recipient for a liquid. Depending on this, the translation will be very
different. But if we see it associated with other terms, as in the following
sentence: “Place the container of dough on the table and pour a cup of water
into the glass,” we will probably find the right translation.
94 François Massion
6 See http://gallantlab.org/huth2016/.
Terminology in the Age of AI 95
Several sciences have contributed over the years to the study of human
cognition: psychology, sociology, neurosciences to name a few. It is now an
established fact that language is embodied, i.e. our senses (sight, hearing, taste,
smell, touch as well as the motor system) play an important role in the way we
understand the meaning of words and sentences (Gallese and Lakoff 2005).
Progresses made in neurosciences allow us to observe via fMRI
(functional magnetic resonance imaging) brain activity in the cortex during
cognition and language tasks. Angela Friederici describes in her seminal
book how the human brain processes language (Friederici and Chomsky
2017) and proposes a language comprehension model. She observes syntac-
tic and semantic processes in the brain, whereas first syntactic processing
occurs and then semantic processing in a different network. This tells us
that human cognition also uses basic rules.
Semantic processes have been studied for example by Friedemann Pul-
vermüller, who observed what he called thought circuits (Pulvermüller 2013;
Pulvermüller, Garagnani, and Wennekers 2014). He could observe learned
neuronal circuits in the brain, e.g. for number words, for action words, for
abstract words. These circuits are developed “by way of associative corre-
lation learning”. He calls them distributed neuronal assemblies (DNA) or
thought circuits. DNAs can be combined into higher-order semantic circuits.
He adds “DNAs also provide a mechanism for combinatorial generalization
and for semantically linking words and concepts (or thoughts)” (Pulvermüller
2013: 585-589).
From these and several associated findings we retain the following
facts that are of some importance when it comes to creating intelligent
terminologies:
1. We have been learning concepts throughout our life mostly via our
senses and our individual experience. This is a reason for which in-
dividuals do not understand things and language in a uniform way
(Dehaene 2018; Frith 2013).
2. Certain concepts have been learned at a very early stage in our life
and are strongly connected to our early sensory and motor experien-
ces. They can be observed in many primitive metaphors like “a warm
person” (Lakoff and Johnson 2008).
96 François Massion
There are several terms used to describe such typical contexts: semantic
network, frame, scenario, mental space …. We will use here the term frame
in this sense.
take for example a neural machine translation system, the learning process
is divided into two main tasks: decoding and encoding. Encoding means
that the neural network takes thousands of source sentences (from bilingual
training material) and learns for each sentence a so-called “thought vector”,
i.e.a series of numbers that reflect the words of the source sentence and their
interdependencies.
After the encoding of the source sentence is completed, the decoding
occurs, looking at the words of the thought vector and trying to predict the
associated words of the target language. The system has the choice between
all the words it has learned and uses the values from the vectors learned
for each individual word in order to calculate the most likely translation.
In the learning phase, the learning algorithm counts how often words like
translation and memory are used together or are used in a similar context.
The result is a model used to predict sequences of words. Thus, when the
decoder has the choice between memory and banana as a term that comes
after the word translation it will take the term memory. It does this using
an output function that looks for a maximum value (softmax) for the combi-
nation of the memory vector and the other word vectors of the vocabulary.
Of course, the learning process is more complex than this simplified
representation. Deep Learning networks that started to be widespread and
successful from 2012 on, use several hidden layers between the input (source
sentence) and the output (machine translation). Intermediate layers are hid-
den and take over specific tasks like, as humans do, looking at the context
of neighbouring words before making a decision on the most appropriate
next term. This is how the so-called CNNs (convolutional neural networks)
operate. The data is passed on from one layer to the next layer. Each layer
is responsible for a specific task. Before being reached out to the next layer,
the numeric data is transformed in order to reflect the intermediary results
(=decisions). In a way this emulates neural networks in the brain that use
neurotransmitters to send information from one neuron to the next.
CNNs have been inspired by human visual perception in the visual
cortex described by David Hunter Hubel and Torsten Nils Wiesel in 1962
and for which they later received the Nobel Prize. They discovered two cate-
gories of cells in the visual cortex, simple and complex cells that processed
information from regions of cortical fields. Complex cells summarized the
stimulus of simple cells and specialized in recognizing certain patterns,
e.g. vertical or horizontal traits. The final image is the result of a sequence
98 François Massion
Currently, the state-of-the-art models used for neural machine translation are
called transformers. Systems that have been trained via Deep learning are
predictive models and do not use rules to generate language or translations.
Because of this they are not always consistent in the use of words and they
make mistakes when words have context-dependent meanings. Advanced
algorithms like BERT, BERTRAM or ELMo are trying to make trained
models more aware of the context but there is still a long way to go as no in-
telligence or reasoning is involved and the data to be computed is immense.
In a recent article titled “Extending Machine Language Models to-
ward Human-Level Language Understanding” (McClelland et al. 2019),
AI researchers suggest enhancing existing machine language models by
building up an artificial fast learning system that stores embeddings derived
from past learning cycles and using them in an integrated system of know-
ledge. In context the system would retrieve relevant information by also
looking at the similarity between the stored embeddings and the context.
The authors write “artificial systems with indefinite capacity could exceed
human abilities in this regard” (McClelland et al. 2019: 7).
Terminology in the Age of AI 99
Some 250 years ago, John Wilkins (Wilkins 1668) dreamed of a universal
language that could describe reality in an unambiguous way to all humans
independently of their own language. He writes:
So that if men should generally consent upon the same way or manner of Ex-
pression, as they agree in the same Notion, we should then be freed from that
Curse in the Confusion of Tongues, with all the unhappy consequences of it.
(Book I, Chap. V, 20)
From the description of human and machine cognition we can now define a
new type of terminologies that can be used by humans and machines alike
and can represent the meaning of terms in context and across languages. We
want to complement the existing multilingual structure of concept-based
terminologies with standard contexts that we call frames. Frames stand for
typical use situations of concepts or terms.
This can be useful not only for homonyms, i.e. clearly different con-
cepts like bridge (related to construction or to a tooth) but also for synonyms
that are not interchangeable all the time like boss and manager or gingivitis
and gum disease.
A particular advantage of this approach concerns multilingual termi-
nologies. On the one hand we may have a concept with a definition shared
equally by all languages like taxi or breakfast, but with different contexts
of use, e.g. a breakfast in France and in China. Relations between each
100 François Massion
language variant and concepts in other languages (e.g. rice vs. croissant in
the case of breakfast) help translators and interpreters in their work.
Another useful application concerns terms and concepts or terms
that have more than one equivalent in the target language. For example,
the English term fish has two translations in Spanish: pescado (= seafood)
and pez, depending on whether the fish is a dish on your plate or is alive.
Each usage situation of the Spanish translation can be linked to a different
concept frame like restaurant, plate, waiter in the one case and sea, water,
swarm in the other case. This is not only very useful for translators, but also
for quality assurance technologies that can automatically interpret these
relations and report a wrong translation in a specific context.
Beside manual research work, tools and methods of natural language proces-
sing (NLP) and artificial intelligence can be used to semi-automatically identify
terms that are used together in the same context. For example, co-occurrence
matrices can be generated and processed to create lists of co-occurring words.
The idea of word embeddings was introduced in 2001 (Bengio, Du-
charme, and Vincent 2001). Word embeddings are learned by machines
and associate a keyword with other words that influence it, have a similar
meaning or behave in the same way. For each word, these associations are
vectors of mostly between 100 and 300 dimensions and can be queried to
get lists of semantically related words, e.g. words related to graduate (see
figure 3). This information is useful for building semantic frames.
• a context for connected relevant terms, e.g. income > equity > taxes
(as opposed to the income > sales > goods, that may require a different
translation in some languages).
• terms with a special usage attribute (e.g. prohibited translation)
• categories of terms based on their properties (e.g. attribute about roles,
named entities, text classification, part of speech, etc.)
An illustration of this are the Wikipedia articles that contain intelligent links
to data associated with the terms highlighted and linked in the articles.8
There are several tools available on the market for text annotation.
Some of them can directly tap intelligent terminology databases.
8 To experience this, you can open a Wikipedia page, e.g. https://en.wikipedia.org/wiki/
Artificial_intelligence, and then click in the left panel under Tools on the menu item:
Wikidata item.
104 François Massion
Annotation can also add markup to content for further processing by diverse
applications. In this way, intelligent applications such as chatbots or smart
assistants can “understand” annotated content, recognize the elements with
relevant information and output the required results. This can already be
seen in areas such as technical support for products or in marketing and
sales, when connected products such as flight, hotel and rental car are offered
to the user as a package.
Quality assurance technologies connected to intelligent terminology
repositories can check context-dependent translation variants and identify
mistranslations in a particular context.
8. Conclusion
Intelligent terminologies are still relatively new, and it will take some time
before they are adopted by translators and interpreters or by academia on a
broad basis. Existing solutions differ in the variety of relations they model
and the methods they use to implement them. Intelligent terminologies are
the result of multidisciplinary research work and are changing the paradigms
of terminology work. Language professionals can benefit from them but will
also discover new services opportunities. With their unique skills, they can
help build up genuinely multicultural and multilingual knowledge bases.
Terminology in the Age of AI 105
Glossary
Expression Description
Concept Abstract unit of thought (e.g. what is understood by ‘software’)
Term Word or group of words denoting a concept
Relation Semantic relation between concepts or terms (e.g. meronomic relation:
steering wheel is part of car)
Frame Typical recurrent context of use for a term or concept.
Bibliographical references
Ogden, Charles Kay, and Ivor Armstrong Richards (1923): The Meaning of
Meaning, London: Routledge & Kegan Paul. <https://books. google.
de/books?id=fo84zQEACAAJ>.
Picht, Heribert, and Klaus-Dirk Schmitz (2001): Terminologie und Wissens-
ordnung: Ausgewählte Schriften aus dem Gesamtwerk von Eugen
Wüster, Term Net. <https://books.google.de/ books?id=utaTAAAA-
CAAJ>.
Pulvermüller, Friedemann (2013) “How neurons make meaning: Brain
mechanisms for embodied and abstract-symbolic semantics”, Trends
in Cognitive Sciences, 17:9, pp. 458-470. <https://doi.org/10.1016/
j.tics.2013.06.004>.
Pulvermüller, Friedemann, Max Garagnani, and Thomas Wennekers (2014)
“Thinking in circuits: Toward neurobiological explanation in cognitive
neuroscience”, Biological Cybernetics, 108:5, pp. 573-93. <https://
doi.org/10.1007/s00422-014-0603-9>.
Quillian, Ross M. (1968) “Semantic memory”, in Semantic Information
Processing: Readings in Cognitive Science. Ed. by Marvin Minsky,
MIT Press.
Quine, Willard V.O. (1969): Ontological Relativity and Other Essays, United
Kingdom, Columbia University Press.
Roche, Christophe (2007) “Le Terme et Le Concept: Fondements d’une
Ontoterminologie”, TOTh 2007: « Terminologie & Ontologie: Théories
et Applications » - Annecy 1er Juin 2007, pp. 1-13.
Roche, Christophe, Marie Calberg-challot, Luc Damas, Philippe Rouard,
Christophe Roche, Marie Calberg-challot, Luc Damas, and Philippe
Rouard Ontoterminology (2011) “Ontoterminology: A new paradigm
for terminology”, Proceedings of the International Conference on
Knowledge Engineering and Ontology Development, Funchal-Ma-
deira, Portugal.
Russell, Bertrand (1919): Introduction to Mathematical Philosophy, London,
George Allen and Unwin.
Temmerman, Rita, and Koen Kerremans (2003) “Termontography: Ontolo-
gy building and the sociocognitive approach to terminology descrip-
tion”, Applied Linguistics, 105, pp. 1-10. <http://citeseerx.ist. psu.edu/
viewdoc/summary?doi=10.1.1.95.3960>.
Wilkins, John (1668): An Essay Towards a Real Character and a Philoso-
phical Language, S. Gellibrand.