Confirmed presentations (updated to 17 August 2017)

* Lists subject to changes

Roberto Navigli

Roberto Navigli is Professor of Computer Science at the Sapienza University of Rome, where he heads the multilingual Natural Language Processing group. He was awarded the Marco Somalvico 2013 AI*IA Prize for the best young researcher in AI. He is one of the few Europeans to have received two prestigious ERC grants in computer science, namely an ERC Starting Grant onmultilingual word sense disambiguation (2011-2016) and an ERC Consolidator Grant on multilingual language- and syntax-independent open-text unified representations (2017-2022).

He was also a co-PI of a Google Focused Research Award on NLP. In 2015 he received the META prize for groundbreaking work in overcoming language barriers withBabelNet, a project also highlighted in TIME magazine and presented in the most cited 2012 paper in the Artificial Intelligence Journal, a journal for which he is currently an Associate Editor.

Based on the success of BabelNet and its multilingual disambiguation technology, he co-founded Babelscape, a Sapienza startup company which enables HLT in hundreds of languages.

KEYNOTE - From text to concepts and back: going multilingual with BabelNet in a step or two

In this talk I will introduce the most recent developments of the BabelNet technology, winner of several scientific awards and of great interest to interpreters and translators. I will first describe BabelNet live – the largest, continuously-updated multilingual encyclopedic dictionary – and then discuss a range of cutting-edge industrial use cases implemented by Babelscape, our Sapienza startup company, including: multilingual interpretation and mapping of terms; multilingual concept and entity extraction from text; semantically-enhanced multilingual translation, and cross-lingual text similarity.

Geoffrey Westgate

Geoffrey Westgate is Head of the Support Section, PCT Translation Division, at WIPO in Geneva, Switzerland. After obtaining a DPhil in 1999 from the University of Oxford, UK, where he also taught German language and literature, he worked initially as a translator and then a reviser in WIPO’s patent translation department. Since 2009 he has headed the Division’s Support Section, with responsibility for computer-assisted translation tools, translation project management, and terminology management, including WIPO’s online terminology portal, WIPO Pearl.

WIPO Pearl - The terminology portal of the World Intellectual Property Organization

In this paper we shall present WIPO Pearl, the multilingual terminology portal of the World Intellectual Property Organization, a specialized agency of the United Nations. The nature of the linguistic dataset made available in WIPO Pearl will be described and we shall show how multilingual knowledge representation is achieved and graphically displayed. Secondly, we shall demonstrate how such data is exploited to facilitate search of prior art for patent filing or examination purposes, by leveraging the validated linguistic content as well as the validated conceptual relations that are presented in “concept maps”. We shall discuss how, in addition to humanly validated concept maps, “concept clouds” are generated by means of machine learning algorithms which automatically cluster concepts in the database by exploiting textual data embedded in the terminology repository. And finally, we shall present opportunities for collaborations with WIPO in the field of terminology.

Carlos Teixeira, Joss Moorkens, Daniel Turner, Joris Vreeke and Andy Way

Carlos Teixeira is a post-doctoral researcher in the ADAPT Centre for Intelligent Digital Content Technology and a member of the Centre for Translation and Textual Studies (CTTS) at Dublin City University (DCU). He holds a PhD in Translation and Intercultural Studies and Bachelor degrees in Electrical Engineering and Linguistics. His research interests include Translation Technology, Translation Process Research, Translator-Computer Interaction, Localisation and Specialised Translation. He has vast experience in the use of eye tracking for assessing the usability of translation tools. His industry experience includes over 15 years working as a translator, localiser and language consultant.

Joss Moorkens is a lecturer in Translation Technology in the School of Applied Language and Intercultural Studies (SALIS) at DCU and a researcher in the ADAPT Centre and CTTS. Within ADAPT, he has contributed to the development of translation tools for both desktop and mobile. He is co-editor of a book on human and machine translation quality and evaluation (due in 2018) and has authored journal articles and book chapters on topics such as translation technology, post-editing of machine translation, human and automatic translation quality evaluation, and ethical issues in translation technology in relation to both machine learning and professional practice.

Daniel Turner is a research engineer in the ADAPT Centre’s Design & Innovation Lab (dLab). Within ADAPT, he has contributed to projects with a strong focus on rapid prototyping of user interfaces. He is proficient in full stack development with experience using a variety of languages and tools.

Joris Vreeke is Scrum Master and Senior Software Engineer in the ADAPT Centre’s dLab. He has a background in software development and design with a preference for graphics, UI/UX and web application development.

Andy Way is a professor in the School of Computing at DCU and leads ADAPT Centre’s Transforming Digital Content theme as well as the Localisation spoke, supervising projects with prominent industry partners. He has published over 350 peer-reviewed papers and successfully graduated numerous PhD and MSc students. His research interests include all areas of machine translation such as statistical MT, example-based MT, neural MT, rule-based MT, hybrid models of MT, MT evaluation and MT teaching.

Creating a tool for multimodal translation and post-editing on touch-screen devices

Only a few translation tools have been created with an ‘organic’ integration of TM and MT, i.e. tools that were designed to work equally well for post-editing MT and for handling TM matches. Still, these scarce options are based on the traditional input modes comprised of keyboard and mouse. Building on our experience in creating a prototype mobile post-editing interface for smartphones, we have created a translation editing environment that accepts additional input modes, such as touch commands (on devices equipped with touch screens, such as tablets and select laptops) and voice commands – using automatic speech recognition. Another important tool feature is the inclusion of accessibility principles from the outset, with the aim of opening translation editing to professionals with special needs. In particular, the tool is being designed to cater for blind translators. Our presentation will report on initial usability tests with an early version of the tool. The results include productivity measurements as well as data collected using satisfaction reports. Our ultimate goal is to test whether the tool can help alleviate some of the pain points of the edit-intensive, mechanical task of desktop post-editing.

Dimitra Kalantzi

Dimitra Kalantzi is a professional English to Greek translator currently based in Athens. Over the past 13 years, Dimitra has worked for companies and translation agencies both in the UK and Greece, as well as for the European Parliament in Luxembourg (as a trainee translator). She has an MSc in Machine Translation from UMIST, UK and a PhD in Informatics (subtitling and linguistics) from the University of Manchester and is a member of the Institute of Translation and Interpreting (ITI) and the European Association for Machine Translation (EAMT).

POSTER - MT and post-editing from a translator’s perspective

There is no doubt that MT is nowadays one of the major trends in the translation industry. Indeed, more and more translation agencies offer MT and post-editing services to their clients, and professional translators are more and more likely to be offered post-editing tasks in their everyday work. In this context, and drawing from my own experience with MT and post-editing as a translator, I will discuss some common myths around MT and post-editing, will suggest some additional services that both translation agencies and freelance translators can offer in relation to MT, and will also put forward some reservations and ideas regarding MT evaluation within the translation industry. This paper will also make a plea to universities and academics involved in the teaching of MT courses and modules to also cater to the needs of practicing translators.

Sheila Castilho, Joss Moorkens, Federico Gaspari, Andy Way, Yota Georgakopoulou, Maria Gialama, Vilelmini Sosoni and Rico Sennrich

Sheila Castilho is a post-doc researcher in the ADAPT Centre in Dublin City University. Her research interests include human and usability evaluation of machine translation, translation technology and audio-visual translation.

 

Co-authors

Joss Moorkens is an Assistant Professor and researcher in the ADAPT Centre, within the School of Applied Languages and Intercultural Studies in Dublin City University (DCU) with interests in human evaluation of translation technology, ethics and translation technology, and translation evaluation.

Federico Gaspari teaches English linguistics and translation studies at the University for Foreigners “Dante Alighieri” of Reggio Calabria (Italy) and is a postdoctoral researcher at the ADAPT Centre in Dublin City University, where he works on EU projects focusing on machine translation evaluation.

Andy Way is a Professor of Computing at Dublin City University (DCU) and Deputy Director of the ADAPT Centre. He is a former President of the European Association for Machine Translation and edits the journal Machine Translation.

Panayota (Yota) Georgakopoulou holds a PhD in translation and subtitling and is a seasoned operations executive in the subtitling and translation industries, with significant experience in the translation academia as well. She is currently Senior Director, Research & Int’l Development at Deluxe Media, leading translation initiatives and research on language technologies and tools, and their application in subtitling workflows.

Maria Gialama is currently working as Account Manager, R&D at Deluxe Media, focusing in the application of language technologies in subtitling. Maria received her MA in translation and subtitling from the University of Surrey and has extensive experience in translation ops management.

Vilelmini Sosoni is Lecturer at the Ionian University in Greece. She has taught Specialised Translation in the UK and Greece and has extensive industrial experience. Her research interests lie in the areas of the translation of institutional texts, translation technology and audiovisual translation.

Rico Sennrich is a research associate at the University of Edinburgh. His main research interest is machine learning, especially in the area of machine translation and natural language processing.

 

Crowdsourcing for NMT Evaluation: Professional Translators versus the Crowd

The use of machine translation (MT) has become widespread in many areas, from household users to the translation and localization industries. Recently, the great interest shown in neural machine translation (NMT) models by the research community has made more detailed evaluation of this new paradigm essential, since several comparative studies using human and automatic evaluation of statistical and neural MT have shown that results concerning the improvements of NMT are not yet conclusive (e.g. Castilho et al. 2017). Crowdsourcing has become a frequently-employed option to evaluate MT output. In the field of translation, such crowds may consist of translation professionals, bilingual fans or amateurs, or a combination thereof. Crowdsourcing activities are at the heart of the European-funded research and innovation project TraMOOC (Translation for Massive Open Online Courses). In this presentation, we will focus on the MT evaluation crowdsourcing activities performed by professional translators and amateur crowd contributors. We will present the results of this evaluation based on automated metrics and post-editing effort and compare how translators and the general crowd assessed the quality of the NMT output.

Jon Riding and Neil Boulton

Jon Riding leads the Glossing Technologies Project for United Bible Societies. The project develops language independent NLP systems to assist Bible translators by automatically analysing elements of natural languages. He is a Visiting Researcher at Oxford Brookes University.
In addition to his work in computational linguistics for UBS Jon teaches Koine Greek, Classical Hebrew and Biblical Studies for various institutions in the UK including Sarum College – (where he is an associate lecturer).
Research interests include the automatic analysis of complex non-concatenative structures in natural language, the development of the New Testament text and the writings of the early Church Fathers.
E: jonriding@biblesocieties.org

 

 

Neil Boulton works as part of the Glossing Technologies Project for United Bible Societies. The project develops language independent NLP systems to assist Bible translators by automatically analysing elements of natural languages. Previously most of his working life has been spent in various IT roles for British and Foreign Bible Society, based in Swindon, UK.
E: neilboulton@biblesocieties.org

(Stand-by for main track) Learning from Sparse Data - Breaking the Big Data Stranglehold

The Bible Societies have for many years built systems to help translators working in scores of languages. Their focus is in linguistic analysis rather than synthesis. But there is a problem, shared by all MT systems. Until there is enough text to train the system the output is limited. By the time there is enough training data much of the task may already be complete.

To address this, United Bible Societies has begun to re-imagine what we expect from our translation support systems. A system is now in development which begins learning about the target language at the very start of a Bible translation project. Rather than building stand alone morphology analysers, glossing engines and aligners the project constructs a learning framework within which all of these machines, and more, can operate with very small amounts of text, using outputs from one context to strengthen a hypothesis from another.

This paper describes a framework within which such processing might take place, how that framework enables learning to take place from very small amounts of data, how that learning is gradually aggregated into a coherent language model and how this model is used to support the translator in their work.

 

Michael Farrell
 

Michael Farrell is an untenured lecturer in computer tools for translators and interpreters at the International University of Languages and Media (IULM), Milan, Italy, the developer of the terminology search tool IntelliWebSearch, a qualified member of the Italian Association of Translators and Interpreters (AITI), and member of the Mediterranean Editors and Translators association.

Besides this, he is also a freelance translator and transcreator. Over the years, he has acquired experience in the cultural tourism field and in transcreating advertising copy and press releases, chiefly for the promotion of technology products. Being a keen amateur cook, he also translates texts on Italian cuisine.

 

Building a custom machine translation engine as part of a postgraduate university course: a case study

In 2015, I was asked to design a postgraduate course on machine translation (MT) and post editing. Following a preliminary theoretical part, the module concentrated on the building and practical use of custom machine translation (CMT) engines. This was a particularly ambitious proposition since it was not certain that students with degrees in languages, translation and interpreting, without particular knowledge of computer science or computational linguistics, would succeed in assembling the necessary corpora and build a CMT engine. This paper looks at how the task was successfully achieved using KantanMT to build the CMT engines and Wordfast Anywhere to convert and align the training data. The course was clearly a success since all students were able to train a working CMT engine and assess its output. The majority agreed their raw CMT engine output was better than Google Translate’s for the kinds of text it was trained for, and better than the raw output (pre-translation) from a translation memory tool. There was some initial scepticism among the students regarding the effective usefulness of MT, but the mood clearly changed at the end of the course with virtually all students agreeing that post-edited MT has a legitimate role to play.

Claire Lemaire and Christian Boitet

Claire Lemaire is a translator who worked in the IT industry before studying computational linguistics. She just finished a PHD on the translation technologies practices of specialised translators and domain experts, in the ILCEA4 laboratory and is currently Visiting Researcher at LIG-GETAL laboratory.

Christian Boitet is emeritus professeor at Université Grenoble Alpes and continues his research on MT and CAT in the LIG-laboratory. He started in 1970 with the Professor Vauquois and succeeded him as director of the GETA study group on MT from 1985 to 2007. His new project aims at using ML (with DL) for developing UNL-based enconverters ans deconverters for as many under-resourced languages as possible.

POSTER - On the Need of New Computer Aids for Translating Writers

Thanks to an experiment in French companies, we found out a new situation of multilingual writing, for which there seems to be no dedicated translation aids. Given the relative lack of competence on the target language (usually but not always English) we propose to build a new kind of CAT tools incorporating interactive (but asynchronous) disambiguation in source at some point during analysis and specific PE aids tuned to the user profiles.

The new system would be a DBMT (Dialogue-Based MT) system, that allows the user to edit the source and target texts, during the translation process, instead of translating and then post-editing the target text. The system would use two advantages of this new situation: (1) the expert knowledge of the “translating writers” on their domain, and (2) the possibility for them to change both texts, unlike a regular translator.

That approach would solve one of the main problems of MT, namely the poor quality of source texts usually given to the translators. Also, the MT results would be immediately assessed and if the translation is bad, immediately produced again after a change in the source text so that the source text be more machine translatable.

Silvia Rodríguez Vázquez and Sharon O'Brien

Dr Silvia Rodríguez Vázquez is a postdoctoral researcher at the Department of Translation Technology (TIM) of the University of Geneva, Switzerland. Silvia’s research interests include web accessibility, localisation, universal design and usability, and the accessibility of translation technologies. Over the last years, she has been a strong advocate for the achievement of an accessible Multilingual Web for all, disseminating her research both through academic and industry-focused publications, including an article on the relevance of considering accessibility best practices in web localisation processes for MultiLingual magazine.

Dr Sharon O’Brien is a senior lecturer in the School of Applied Language and Intercultural Studies, Dublin City University, Ireland. She is also a Funded Investigator in the Science Foundation Ireland funded research centre, ADAPT, and was Director of the Centre for Translation and Textual Studies at DCU. Her research interests include translator-computer interaction, localisation, cognitive ergonomics in translation and translation quality assessment. She previously worked in the localisation sector as a language technology specialist.

POSTER - Web accessibility compliance in localisation: the missing link for an optimal end user experience

In an increasingly competitive business landscape, the ever-evolving localisation industry is now striving for differentiation. One of the strategies adopted, particularly by the largest multinationals, has been to expand their service coverage beyond traditional localisation and the provision of translation technology to satisfy new digital market needs. Concretely, we have observed a considerable increase in the number of companies showcasing knowledge and know-how in Digital Marketing and User Experience Design, always with a clear goal: enhancing the final end user experience when interacting with multilingual web content. But are we really ensuring an optimal experience for all?

If the localisation industry is looking to consolidate this strengthened service portfolio, awareness of key human-computer interaction aspects and best practices, including web accessibility standards, could be crucial for success. Drawing upon the data collected through a series of interviews with representatives of six world-leading multinational companies from the localisation industry and one of their clients, this paper will report on the readiness of current localisation workflows and professionals to deliver more accessible multilingual websites for all. We will also review the overlaps between responsive design, SEO and current web accessibility guidelines, and present how their compliance could bring competitive advantage to localisation businesses.

Emmanuelle Esperanca-Rodier, Laurent Besacier and Caroline Rossi

Emmanuelle Esperança-Rodier is a lecturer at Univ. Grenoble Alpes (UGA), France, where she teaches English for Specific Purpose and a member of the Laboratoire d’Informatique de Grenoble (LIG). After defending a PhD in computational linguistics, on “Création d’un diagnostic générique de langues contrôlées, avec application particulière à l’anglais simplifié”, she worked as a post-editor in a translation agency. Back at University, she participated in IWSLT and WMT evaluation campaigns, as well as in several LIG projects. She now works on the evaluation of MT systems based on competences and focused on tasks, translation error analysis and multilinguism.

Co-authors

Prof. Laurent Besacier defended his PhD thesis (Univ. Avignon, France) in Computer Science in 1998 on “A parallel model for automatic speaker recognition”. Then he spent one and a half year at the Institute of Microengineering (EPFL, Neuchatel site, Switzerland) as an associate researcher working on multimodal person authentication (M2VTS European project). Since 1999 he is an associate professor (full professor since 2009) in Computer Science at Univ. Grenoble Alpes (he was formerly at U. Joseph Fourier). From September 2005 to October 2006, he was an invited scientist at IBM Watson Research Center (NY, USA) working on Speech to Speech Translation.
His research interests are mainly related to multilingual speech recognition and machine translation. Laurent Besacier has published 200 papers in conferences and journals related to speech and language processing. He supervised or co-supervised 20 PhDs and 30 Masters. He has been involved in several national and international projects as well as several evaluation campaigns. Since October 2012, Laurent Besacier is a junior member of the “Institut Universitaire de France” with a project entitled “From under-resourced languages processing to machine translation: an ecological approach”.

Caroline Rossi is a lecturer in the Applied Modern Languages department at Univ. Grenoble Alpes, where she teaches English and translation. She is a member of the Multilingual Research Group on Specialized Translation (GREMUTS) within ILCEA4 (Institut des Langues et Cultures d’Europe, Amérique, Afrique, Asie, Australie). Her current research focus is on integrating critical skills and understanding of both statistical and neural machine translation in translator training.

Since 2014, Alexandre Bérard has been a PhD Student at the University of Lille (with Prof. Laurent Besacier, and Prof. Olivier Pietquin). He worked with the SequeL team (specialized on Machine Learning) at Inria Lille, and then from 2016, with GETALP (specialized on NLP) at the University of Grenoble.
Specialized in Neural Machine Translation techniques, in particular for automatic post-editing and end-to-end speech translation, he obtained a Software Engineering degree from the INSA of Rennes in 2014, and a Master’s degree in data science from the University of Rennes.

Evaluation of NMT and SMT systems : A study on uses and perceptions

Statistical and neural approaches have permitted fast improvement in the quality of machine translation, but we are yet to discover how those technologies can best “serve translators and end users of translations” (Kenny, 2017). To address human issues in machine translation, we propose an interdisciplinary approach linking Translation Studies, Natural Language Processing and Philosophy of Cognition. Our collaborative project is a first step in connecting sound knowledge of Machine Translation (MT) systems to a reflection on their implications for the translator. It focuses on the most recent Statistical MT (SMT) and Neural MT (NMT) systems, and their impact on the translator’s activity. BTEC-corpus machine translations, from in-house SMT and NMT systems, are subjected to a comparative quantitative analysis, based on BLEU, TER (Translation Edit Rate) and the modified version of METEOR from the LIG (Servan & al, 2016). Then, we qualitatively analyse translation errors from linguistic criteria (Vilar, 2006) or the MQM (Multidimensional Quality Metrics) using LIG tools, to determine for each MT systems, which syntactic patterns imply translation errors and which error type is mainly made. We finally assess translators’ interactions with the main error types in a short post-editing task, completed by 10 freelance translators and 20 trainees.

Corey Miller, Danielle Silverman, Vanesa Jurica, Elizabeth Richerson and Rodney Morris

Corey Miller is a linguist working at MITRE Corporation in USA. Trained in sociolinguistics and phonetics, he has worked for several years in the speech technology industry. More recently he has been focusing on problems in computer-assisted translation and terminology management, where he hopes to eventually incorporate his interests in sociolinguistics and speech.

(Stand-by for main track) Enlisting CAT-tools in the battle for register equivalence

There seems to be general agreement that aiming for equivalence in register between source and target languages in translation is a good thing; in fact, this is now encoded in translation services standard ISO 17100:2015 as “compliance with any proprietary and/or client style guide (including register and language variants)”. The question this paper will address is what specific things computer-assisted translation (CAT) tools can do to aid in this process. We will discuss the role of both termbases and translation memories (TMs) in recording register information. Finally, we will discuss work on register in machine translation (MT) and how this may ultimately facilitate a register-enabled CAT workflow.

It is hoped that this discussion specifying ways in which the CAT workflow can be fortified to accommodate register information will provide researchers and developers a path forward for a forthcoming generation of CAT tools which will make it easier for translators and reviewers to maintain and assess register fidelity.

Samuel Läubli and David Orrego-Carmona

Samuel Läubli
I am a PhD Student in Machine Translation at the University of Zurich (CH). I obtained a BA in Computational Linguistics and Language Technology from the University of Zurich (CH) in 2012, and a Master of Science in Artificial Intelligence from the University of Edinburgh (UK) in 2014. From 2014 to 2016, I implemented machine translation systems for post-editing as a Senior Computational Linguist at Autodesk, Inc. My research focus is the intersection of Machine Translation, Translation Process Research, and Human–Computer Interaction.

David Orrego-Carmona
I am a Lecturer in Translation Studies at Aston University (UK) and an associate research fellow at the University of the Free State (SA). After completing a BA in Translation at the Universidad de Antioquia (Colombia) and working as an in-house translator, I gained an MA (2011) and a PhD (2015) in Translation and Intercultural Studies from the Universitat Rovira i Virgili in Tarragona (Spain). In 2016 I was a post-doctoral research fellow at the University of the Free State. My areas of expertise include audiovisual translation and translation technologies. In particular, my research analyses how translation technologies empower users and how the democratization of technology allows these users to become translators.
Other interests include translation process research and the cognitive exploration of translation production and reception, mainly using eye tracking technologies.

When Google Translate is better than some human colleagues, those people are no longer colleagues

Expressing discomfort with machine translation is a recurrent theme in translator forums on social media. We analyse posts from Facebook, LinkedIn, and Twitter, showing that translators spend considerable time and effort on mocking and discussing machine translation failures. Research has shown a disconnect between translation technology researchers and practitioners, while at the same time indicating the benefits of collaboration between them. Reflecting on an analysis of posts and engagement of translators in social media, we outline three suggestions to bridge this gap: (i) identify and report patterns rather than isolated errors, (ii) organise or participate in evaluation campaigns, and (iii) engage in cross-disciplinary discourse. Rather than pointing out each other’s deficiencies, we call for computer scientists, translation scholars, and professional translators to advance translation technology by acting in concert.

Joss Moorkens

Joss Moorkens is a lecturer in Translation Technology in the School of Applied Language and Intercultural Studies in Dublin City University (DCU) and a researcher in the ADAPT Centre and the Centre for Translation and Textual Studies. Within ADAPT, he has contributed to the development of translation tools for both desktop and mobile. He is co-editor of a book on human and machine translation quality and evaluation (due in 2018) and has authored journal articles and book chapters on topics such as translation technology, post-editing of machine translation, human and automatic translation quality evaluation, and ethical issues in translation technology in relation to both machine learning and professional practice.

(Stand-by for main track) A Survey of Accredited Irish Language Translators

Translators are currently facing changes in employment conditions and work practices as the technologised ‘vendor model’ has become commonplace. The situation for Irish language translators is particularly dynamic. Language technology is belatedly making a real impact, fuelled by regulatory compliance needs and the removal of the derogation of Irish as an official language of the EU. This study surveys almost half of the 178 accredited Irish language translators, focusing on their opinions on the role of a translator, use of translation technology, the status of Irish within the EU, and job satisfaction. Their responses contrast the perspectives of freelance and public service translators. Participants in this survey are passionate about translation and about the Irish language, and while some translators appear to be thriving, many are not. Responses show a fear of dropping translation quality, a lack of translation talent, and the lack of an audience for translation. Participants’ perspectives differ based on their conditions of employment, with disparities apparent in ratings of job satisfaction – particularly regarding their sense of purpose and payment terms, and in their expressions of positive or negative prospects for the future of their profession.

Martin Kappus, Erik Angelone, Romina Schaub-Torsello, Martin Schuler and Maureen Ehrensberger-Dow

Martin Kappus is a lecturer in the ZHAW Institute of Translation and Interpreting. Before joining the ZHAW faculty, he worked for a CAT tool manufacturer and a large language service provider. His research and teaching interests are language technology in general, translation technology in particular, and barrier-free communication.

Martin Schuler is a research associate and head of the usability lab at the ZHAW School of Applied Linguistics. He has a BA in technical communication and an MA in Human Computer Interaction Design. He has been involved in several types of usability projects for a variety of clients.

Co-authors

Maureen Ehrensberger-Dow is professor of translation studies in the ZHAW Institute of Translation and Interpreting. She has been the (co)investigator in several interdisciplinary projects investigating the reality and ergonomics of professional translation.

Romina Schaub-Torsello is a research assistant at the ZHAW Institute of Translation and Interpreting. She is a trained translator and graduated from the ZHAW MA program in Applied Linguistics with a thesis about the impact that disturbances can have on translator’s cognitive flow and behaviour.

Erik Angelone is professor of translation studies at the ZHAW Institute of Translation and Interpreting. His research interests are in process-oriented translator training, translation pedagogy and curricular design, and empirical translation studies.

 

 

 

When is less actually more? A usability comparison of two CAT interfaces

The body of evidence is growing that CAT tools have fundamentally altered the tasks that most non-literary translators engage in and possibly also their cognitive processing. Recent research suggests that translators may be exposing themselves to unnecessary cognitive friction by the way they use their tools (O’Brien et al. 2017). If tool settings and features do not align with translators’ ways of working, then flow can be interrupted and cognitive load increased. Fatigue and reduced attention are two consequences of cognitive overload over extended periods, both of which have been associated with an increase in errors and lower productivity. We report on a usability comparison of two interfaces for translation work that differ with respect to the information and functions available on the screen when the factory default settings are used (i.e. one interface has several fields with supporting functions visible and the other has a simpler look). Eye tracking measures and indicators from retrospective commentaries and interviews highlight how novices interact with the two interfaces and various features. We consider the implications of our findings in light of recent calls for less cluttered user interfaces and open the discussion of how cognitive load can be reduced.

Andrzej Zydroń

Andrzej Zydroń MBCS CITP

CTO @ XTM International, Andrzej Zydroń is one of the leading IT experts on Localization and related Open Standards. Zydroń sits/has sat on, the following Open Standard Technical Committees:

1. LISA OSCAR GMX
2. LISA OSCAR xml:tm
3. LISA OSCAR TBX
4. W3C ITS
5. OASIS XLIFF
6. OASIS Translation Web Services
7. OASIS DITA Translation
8. OASIS OAXAL
9. ETSI LIS
10. DITA Localization
11. Interoperability Now!
12. Linport

Zydroń has been responsible for the architecture of the essential word and character count GMX-V (Global Information Management Metrics eXchange) standard, as well as the revolutionary xml:tm (XML based text memory) standard which will change the way in which we view and use translation memory. Zydroń is also chair of the OASIS OAXAL (Open Architecture for XML Authoring and Localization) reference architecture technical committee which provides an automated environment for authoring and localization based on Open Standards.
Zydroń has worked in IT since 1976 and has been responsible for major successful projects at Xerox, SDL, Oxford University Press, Ford of Europe, DocZone and Lingo24 in the fields of document imaging, dictionary systems and localization. Zydroń is currently working on new advances in localization technology based on XML and linguistic methodology.
Highlights of his career include:
1. The design and architecture of the European Patent Office patent data capture system for Xerox Business Services.
2. Writing a system for the automated optimal typographical formatting of generically encoded tables (1989).
3. The design and architecture of the Xerox Language Services XTM translation memory system.
4. Writing the XML and SGML filters for SDL International’s SDLX Translation Suite.
5. Assisting the Oxford University Press, the British Council and Oxford University in work on the New Dictionary of the National Biography.
6. Design and architecture of Ford’s revolutionary CMS Localization system and workflow.
7. Technical Architect of XTM International’s revolutionary Cloud based CAT and translation workflow system: XTM.

Specific areas of specialization:
1. Advanced automated localization workflow
2. Author memory
3. Controlled authoring
4. Advanced Translation memory systems
5. Terminology extraction
6. Terminology Management
7. Translation Related Web Services
8. XML based systems
9. Web 2.0 Translation related technology

Beyond Neural MT

A lot of hype and excitement has surrounded the latest advances in Neural Machine Translation (NMT) and generally with some justification: output is generally more fluent and closer to normal human output. Nevertheless some of the claims need to be qualified and practical implementation of NMT is not without difficulty: typically double the training material is required when compared to Statistical Machine Translation (SMT) and training a new engine can take weeks not days. Although NMT can produce much more fluent output than SMT it can have limited impact concerning real world localization tasks. Extensive tests have proven that in the end there is no great improvement in post-editing throughput and NMT is no panacea for omission or mistranslation. With NMT it is also impossible to ‘tune’ the output as can be done with SMT: you have no knowledge of how the NMT engine has made its decisions.

Most practical translation projects do not have anywhere near enough training data and do not have the luxury of waiting weeks for an engine to be trained even if there was enough data. If the training has been done on unrelated material or material that is not directly relevant to the customer’s terminology then misleading results can be produced: due to its improved fluency NMT can make identifying mistakes more difficult. NMT quality can also drop with the length of sentences and translating from a morphologically rich language to one that is morphologically impoverished can in fact produce worse results than SMT.

Marianne Starlander

Dr. Marianne Starlander is a CAT tool specialist and lecturer at the Faculty of Translation and Interpreting of the University of Geneva. She joinded the multilingual information processing department in 2000 where she worked as a teaching and research assistant and now as teaching staff. She originally trained as a translator at the same faculty and also holds a post-graduate degree in European studies from the European Institute of the Unversity of Geneva (2000).
She was coordinator for SUISSETRA (Swiss association for the promotion of CAT tools) from 2008-2012. She has been involved in the research project MedSLT (Medical spoken language translator) and worked on spoken language translation evaluation issues in the frame of her thesis published in 2016. She is responsible for the CAT tools training at the MA level and involved in the continuous education program in CAT tools.

Adapting a Computer Assisted Translation MA Course to New Trends

The present paper will present how we adapted our MA CAT tool class to two current trends. The first trend is the drastic rise of the number of students enrolled in the class. We report on the impact of the rise of students by presenting how challenging it has become to give them the assignment described in Starlander et Morado Vazquez (2013) and in Starlander (2015) The discussion is oriented towards how to teach this evaluation methodology in a different way, by adapting the content and most of all the teaching methods (crowdsourcing, online quiz). We describe in detail how these new activities fit into the entire course content. The second trend is the integration of MT into CAT tools. How can we best introduce this evolution in our teaching? We present the main results of a preliminary experience of integrating a translation exercise involving the use of MT. The final discussion is dedicated to more general teaching challenges implied by the ever moving trends in teaching translation technology.

Claudio Fantinuoli

Claudio Fantinuoli is Lecturer at the Johannes Gutenberg University Mainz in Germersheim and at the Institute for Translation Studies in Innsbruck. His research and teaching areas are Language Technologies in Translation and Interpreting.

Speech Recognition in the Interpreter Workstation

In recent years, computer-assisted interpreting (CAI) programs have been used by professional interpreters to prepare assignments, to organize terminological information, and to share event-related information among colleagues (Fantinuoli, 2016, 2017).

One of the key features of such tools is the ability to support users in accessing terminology during simultaneous interpretation (SI). With state-of-the-art CAI tools, interpreters need to manually input a term or part of it in order to query the database. The main drawback to this approach is that from the booth it is considered both time-consuming and, to some extent, distracting during an activity that requires concentration and rapid information processing (Tripepi Winteringham, 2010). However, initial empirical studies on the use of such tools seem to support the idea that interpreters on the job may have the time and the cognitive ability to look up terms. Furthermore, CAI tools seem to contribute to improving terminology and overall interpreting performance (Prandi 2015; Biagini 2016). With this in mind, the automatization of the querying system would represent a step forward in reducing the additional cognitive effort needed to perform this human-machine interaction. With more free cognitive ability at disposal, it is reasonable to assume that a CAI tool equipped with an automatic look up system would contribute to further improving the terminology and overall performance of interpreters during the simultaneous interpretation of specialized texts.

Speech Recognition (SR) has been proposed as a methodology and technology to automatize the querying system of CAI tools (Fantinuoli 2016; Hansen-Schirra 2012). In the past, the difficulty to build SR systems that were accurate enough to be useful outside of a carefully controlled environment has hindered its deployment in the interpreting setting. However, recent advances in Artificial Intelligence, especially since the dissemination of deep learning and neural networks, have considerably increased the quality of SR (Yu and Deng, 2015). In order to be successfully integrated in an interpreter workstation, both SR and CAI tools must fulfil a series of specific requirements. For example, SR must be truly speaker-independent, have a short reaction time, and be accurate in the recognition of specialized vocabulary. On the other hand, CAI tools need to overcome some shortcomings of current implementations and need, for instance, to handle morphological variants of the selection of results and new ways to present extracted terminology.

In the first part of the paper, a framework for the integration of SR in CAI tools will be defined. In particular, much attention will be devoted to the analysis of state-of-the-art SR and the problems that may arise with its integration into an interpreter workstation. Secondly, the adaptation of traditional querying systems used in CAI tools to allow for keyword spotting will be discussed and a prototype will be presented. Finally, general advantages and shortcomings of SR-CAI integration will be highlighted and prospective developments of the use of SR that support the access of terminological data will be introduced, i.e. the recognition of numbers and entities.

Chong Zhang and Matteo Capelletti

Chong Zhang is a PhD candidate at the Linguistics Department of Stony Brook University
(USA), and works at Lionbridge Inc. Tempere, Finland as an analytical linguist.
Matteo Capelletti is a Project Manager at Lionbridge Inc. Tempere, Finland. He is in charge
of the Social Sentiment Index project at Lionbridge.

POSTER - A Case Study of Machine Translation in Financial Sentiment Analysis

The Social Sentiment Index (SSIX) project [1] aims at overcoming language barriers and realizing a financial sentiment platform capable of scoring textual data in different languages. Since the cost of multilingual sentiment analysis can be high, our project deploys a relatively low-cost yet sufficient approach to help small-scale companies expand to new markets through analyzing social contents in any languages. Our approach first built a sentiment gold standard (GS) corpus (e.g. English) annotated by human financial domain experts. Then the GS corpus was translated into a target corpus (e.g. German) by three machine translation (MT) engines (Microsoft, Google, and Google Neural Network) integrated into Geofluent [2] to allow pre-/post-processing, such as DO-NOT-TRANSLATE rules to tackle with special financial terms.

Our approach takes the maximal advantage of human translation while keeping the cost low by incorporating MT. A crucial prerequisite is that the sentiment of a statement in the GS corpus can be transferred to the target corpus after translation. To meet the prerequisite, we investigate the impact of MT on the sentiment quality of the GS. Experiment 1 A sample of 700 English tweets was selected. It was translated into German simultaneously by one human translator and three MT engines. Then BLEU [3] was calculated to evaluate the quality of the translations. Results suggested that Google and Google Neural Network performed better than Microsoft on 1-gram, and vice versa on the rest.

Engine Microsoft Google Google Neural Network 1-gram 2-grams 3-grams 4-grams 0.901470798 0.865873923 0.786125067 0.684824095 0.963509145 0.846959705 0.728174371 0.605465403 0.963340387 0.846025029 0.727096883 0.604167208

For the purpose of SSIX, the quality of translation is considered the best when there is minimal discrepancy in sentiment between the original and the translation. To explore the sentiment preservation, we conducted Experiment 2.

Experiment 2 A subset of the sample (N = 200) was selected and three German financial domain experts assigned the translations sentiment scores. The regression analyses were done with the lmer4.0 package in R.

We first compared the sentiment of the English GS with the four translations. Results showed that the three MTs had significantly different sentiment from the GS (Microsoft: t = -3.574, p < .001; Google: t = 2.038, p < .05; Google Neural Network: t = 3.101, p < .01), but human translation did not have significantly different sentiment from the GS. Thus we conclude that human translation is the most reliable in sentiment preservation. Then, using human translation as the benchmark, we compared it to MTs. Results showed Microsoft did not preserve the sentiment (t = -2.16, p < 0.05), but both Google and Google Neural Network did (no significant difference). Conclusion Our research provides evidence that sentiment transfer can be successful through MT if using Google and Google Neural Network in Geofluent. This is a crucial step towards achieving a multilingual sentiment platform in the domain of finance. Next, we plan to integrate language-specific processing rules to enhance the performance of MT.

Bianca Prandi

Bianca Prandi is a doctoral student at the University of Mainz/Germersheim. She holds a BA in Intercultural Linguistic Mediation and a MA in Interpreting from the University of Bologna/Forlì. She graduated with a dissertation on the integration of the CAI tool InterpretBank in the curriculum of interpreting students. She is currently working on her doctoral dissertation at Mainz University under the supervision of Prof. Dr. Hansen-Schirra. Her main research interests are new technologies in interpreting and cognition.

CAI tools in the interpreting booth: a proposal for a theoretical model and research methodology

Even though studies on computer-assisted interpreting still represent a very small percentage in the body of research, the topic is starting to gain attention in the interpreting community. So far, only a handful of studies have focused on the use of CAI tools in the interpreting booth (Gacek, 2015; Biagini, 2015; Prandi, 2015a, 2015b). While they did shed some light on the usability and the reception of CAI tools as well as on the terminological quality of simultaneous interpreting performed with the support of such tools, these studies were only product-oriented. We still lack process-oriented, empirical research on computer-aided interpreting. A pilot study currently underway at the University of Mainz/Germersheim (Prandi, 2016, 2017) aims at bridging this gap by combining process- and product-oriented methods. After discussing the theoretical models adopted to date in CAI research, this paper will suggest how an adaptation of Seeber’s (2011) Cognitive Load Model can be better suited then Gile’s (1988, 1997, 1999) Effort Model to operationalize hypotheses on the use of CAI tools in the booth. The paper will then introduce the experimental design adopted in the study with a focus on the features of the texts used and on the rationale behind their creation.

Bert Wylin and Hendrik Kockaert

Bert Wylin (MA Applied Linguistics – Roman languages) has both an academic and a business profile. He works at the K.U.Leuven since 1993, doing research in the fields of technology supported teaching and (language) learning. In 2001, he founded a spin-off company, now merged into Televic Education, developing and servicing educational e-learning and e-assessment projects worldwide and specializing in languages and computer assisted language learning (CALL).

Building Artificial Intelligence on top of a Translation Waste Mountain

The Translation department of the KU Leuven researches the revision and correction of translations. Recently we joined forces with Televic to build our own tools for smart translation revision: we are interested in the “waste”, in the errors that translators make. From a didactical point of view, the analysis of those errors is as interesting as the correct translation. We will show which (game changing) teaching and learning conclusions we can draw from the analysis of the “waste mountain”. TranslationQ and RevisionQ are two tools to evaluate and score translations. Translation evaluation is an important and labour intensive task in the training and in the selection of good translators. Mostly this work is done by human evaluators and has to be repeated for every single translation. Our academic experiments have proven that both tools are as accurate and even more objective than a human evaluation. TranslationQ and RevisionQ are especially useful to evaluate large groups of candidates. Finally, the language correction algorithms have been developed to be language independent, making the tools useable for many language combinations.

Koen Kerremans and Helene Stengers

Koen Kerremans is professor in terminology, specialised translation and translation technology at the department of Linguistics and Literary Studies of Vrije Universiteit Brussel, where he obtained his PhD in 2014. His research interests pertain to applied linguistics, language technologies, ontologies, terminology, special language and translation studies. He is the coordinator of VUB’s master programme of translation and teaches courses on terminology, technical translation and technologies for translators in the master programmes of translation and interpreting.

Helene Stengers is professor in Spanish proficiency, translation and Interpreting at the department of Linguistics and Literary Studies of Vrije Universiteit Brussel, where she obtained her PhD in 2009. Her research interests lie in applied comparative linguistics (especially English and Spanish), cognitive linguistics, phraseology and Foreign Language acquisition (mainly vocabulary acquisition) from a multilingual and intercultural perspective with a view to optimize Foreign Langage pedagogy, as well as translation and interpreter training. She is the research director of the Brussels Institute for Applied Linguistics.

POSTER - Using online and/or mobile virtual communication tools in interpreter and translator training: pedagogical advantages and drawbacks

In this paper we discuss some preliminary results of a comparative study into the use of online and/or mobile virtual communication tools in the master programmes of interpreting and translation at Vrije Universiteit Brussel. Both master programmes are based on a situated learning model, which is generally understood as a didactic method in which translation and interpreting students learn the profession and acquire professional skills through hands on experience by exposing them to simulated or real work environments, situations and tasks. In recent years, this learning-by-doing approach (or authentic experiential learning) has gained quite some traction in translator and interpreter education. In creating authentic learning contexts for student translators and interpreters, technology has become an important factor to take into consideration, given the unmistakable impact that it has on professional translation and interpreting practices. After a review of previous studies dealing with the use of virtual technologies in translator and interpreter training, several virtual communication tools will be tested and evaluated both from the trainers’ and the trainees’ perspectives. Finally, we will reflect on the tools’ pedagogical advantages and drawbacks in order to formulate recommendations for using these technologies in translator and interpreter training contexts.

Laura Cacheiro Quintas

Laura Cacheiro Quintas is a PHD Student in Audiovisual Translation and new technologies. Her thesis is co-directed by the Université de Perpignan Via Domitia (France) and the Universitat Jaume I (Spain). Her research focuses on the integration of new technologies in the subtitling of video lectures with the purpose of assisting translators in their professional activity.
Currently working as a translator for the MIRO Programme at the University of Perpignan, where she tests the implementation of CAT tools and their adaptation to translation workflows, Laura also teaches Audiovisual Translation and Interpreting to second and third year undergraduate students at this university.

POSTER - Towards a Hybrid Intralinguistic Subtitling Tool: MIRO Translate

Making audiovisual educational material accessible for non-native speakers and people who are deaf or hard-of-hearing is an ongoing challenge and the state-of-the-art in this field shows that no current software provides a fully automatic, high-quality solution. This article presents Miro Translate, a hybrid intralinguistic subtitling tool developed in response to this challenge by the MIRO Programme at the University of Perpignan Via Domitia. This cloud-based solution proposes synergies between machine and human subtitlers to simplify the technical and linguistic tasks of subtitlers. To this end, Miro Translate integrates the automatic speech recognition (ASR) technology provided by the Microsoft Translator Speech API, a deep neural network (DNN) system and TrueText technology to improve speech readability. Auto-generated captions constitute a solution to enhance productivity. Nevertheless, certain subtitling conventions must be considered to provide a readable and legible target output. A set of pre-editing and post-editing functionalities have been identified to assist the machine in the spotting task and to improve the quality of auto-generated captions. In conclusion, Miro Translate constitutes a cost-efficient solution to meet the increasing demand of high quality captions for video lecturers.

Kristiina Abdallah

Kristiina Abdallah has worked as a translator, subtitler and technical writer. Since 2001 she has held various positions at the University of Tampere, namely that of an assistant, a lecturer and researcher. As of 2010 she has worked as a university teacher, first at the University of Eastern Finland, and currently at the Universities of Vaasa and Jyväskylä. She defended her doctoral thesis entitled Translators in Production Networks. Reflections on Agency, Quality and Ethics in 2012. Her research interests are translation sociology and, more specifically, translators’ workplace studies.

Three-dimensional quality model: The focal point of workflow management in organisational ergonomics

Although quality is a central concept in every act of translating, it has been considered difficult to define and therefore remained elusive. Generally, approaches to quality, both in translation studies and in translation industry, have concentrated on the product and/or process quality. Yet, in the present-day man and machine mediated, collaborative translation production networks, the challenge to define and manage quality comprehensively has become more acute than ever before.

This paper participates in the discussion on organizational ergonomics of translation by presenting a three-dimensional quality model. The model encompasses not only the so-far familiar product and process dimensions, but also a third dimension called social quality. Social quality, the focus of this paper, addresses the relations of the actors involved, both human and non-human, and their organizational interaction. The theoretical discussion on quality will be complemented by a recent case from Finland regarding the working conditions of the audio-visual translators of Star Wars: The Force Awakens and their impact on translation quality. By emphasizing the point that quality is a multidimensional concept which also includes social and ethical aspects, the paper argues for workflow management that caters to the needs of people, who are the bedrock of the industry.

Vincent Vandeghinste, Sven Coppers, Jan Van den Bergh, Tom Vanallemeersch, Els Lefever, Iulianna van der Lek-Ciudin, Kris Luyten and Frieda Steurs

Dr. Vincent Vandeghinste is a post-doctoral researcher at the KU Leuven, and has been working on natural language processing and translation technologies since 1998. He is the project coordinator of the SCATE project (Smart Computer-Aided Translation Environment), and (co)-authored about 70 publications in the areas of corpus building, treebanking, machine translation, augmented alternative communication and text-to-pictograph translation. He teaches Natural Language Processing, Language Engineering Applications and Linguistics And Artificial Intelligence in the advanced masters program for Artificial Intelligence at KU Leuven, as well as Computational Linguistics to students of Linguistics.Vincent Vandeghinste is a post-doc researcher at the University of Leuven and has been working on machine translation since 2004. He is the project coordinator of the SCATE project, a 3 million euro Flemish project to improve the translation environment of professional translators.

Sven Coppers studied computer science at Hasselt University (UHasselt) and is interested in various aspects of Human Computer Interaction, such as 2D and 3D visualizations, user-centered software engineering, context-awareness and intelligibility (comprehensibility). Currently, he is doing a PhD about making context-aware Internet-of-things applications more understandable and controllable for end-users. In addition, he is working on user interfaces for translation environment within the SCATE project, with a focus on usability, intelligibility, customization and collaboration.

Dr. Jan Van den Bergh is a post-doctoral researcher and research assistant at Hasselt University and member of the HCI group in the research institute Expertise Centre for Digital Media. His research is situated in user-centred engineering of context-aware, mobile or collaborative systems. His recent research is focused on how interactive technology can support knowledge workers and/or end users in specific domains, including professional translation and human-robot collaboration in manufacturing. He obtained a PhD in computer science (human-computer interaction) from Hasselt University. He co-organized several scientific workshops and served as PC member for several conferences and workshops. He is a member of the IFIP working group 13.2 on User-Centred Systems Design.

Tom Vanallemeersch is a researcher in the field of translation technology at KU Leuven. After his studies in translation and in language technology during the early nineties, he focused his attention on various forms of translation software, including translation memories, automated alignment and machine translation. His career involves both academia and industry. He worked in two Belgian translation agencies (Xplanation, LNE International), a French MT development company (Systran), and the MT team of the Commission’s DG Translation in Luxembourg. In academia, he taught the ins and outs of TM and MT to Applied Linguistics students at Lessius Hogeschool (now KU Leuven), then started working at the University’s Centre for Computational Linguistics. He currently performs research in SCATE (Smart Computer-Aided Translation Environment), an extensive, six-team project coordinated by the Centre. While Tom is passionate about translation technology, his career sporadically shifted to other types of natural language processing, such as terminology extraction (coordination of the TermTreffer project at Dutch Language Union).

Dr. Els Lefever is an assistant professor at the LT3 language and translation technology team at Ghent University. She started her career as a computational linguist at the R&D-department of Lernout & Hauspie Speech products. She holds a PhD in computer science from Ghent University on ParaSense: Parallel Corpora for Word Sense Disambiguation (2012). She has a strong expertise in machine learning of natural language and multilingual natural language processing, with a special interest for computational semantics, cross-lingual word sense disambiguation, event extraction and multilingual terminology extraction. She is currently involved in the SCATE project (work package on bilingual terminology extraction from comparable corpora) and the Multilingual IsA project (multilingual database of hypernym relations) and supervises PhD projects on terminology extraction from comparable corpora, semantic operability of medical terminology, irony detection and disambiguation of terminology in a cross-disciplinary context. She teaches Terminology and Translation Technology, Language Technology and Digital Humanities courses.

Ayla Rigouts Terryn is a PhD researcher at the Language and Translation Technology Team (LT3) research group, at the Department of Translation, Interpreting and Communication of Ghent University. She graduated from the University of Antwerp in 2014 with a Master’s in Translation and worked there as a scientific fellow on a one-year research project about translation revision competence. In 2015, she joined the LT3 research group to work on the terminology work package of the SCATE project. She is currently working as an FWO scholar on her PhD about bilingual terminology extraction from comparable corpora.

Bram Bulté obtained an MA in linguistics and literature (2005) and a PhD in linguistics (2013) from Brussels University, and an MA in statistics (2016) and in artificial intelligence, option speech and language technology (2017) from KU Leuven. He worked as a translator for the European Parliament (2007-2015) and as a guest professor at Brussels University (2015-2017). He currently works for the Centre for Computational Linguistics at KU Leuven. His research focuses on second language acquisition, multilingual education and natural language processing.

Iulianna van der Lek is passionate about Language Technologies, always looking for ways to improve the translators’ efficiency. She is currently working as a Research Associate and teaching assistant at KU Leuven, Faculty of Arts, Campus Antwerp. Her research focuses on computer-assisted translation tools and their impact on the translation process, usability, and methods of acquiring domain-specific terminology. As a certified memoQ, Memsource and SDL trainer, she is teaching computer-assisted translation tools both to students and freelance translators. Besides research and teaching activities, she is also coordinating the Postgraduate Programme in Specialised Translation, developing new modules on language technologies and training programs for professional translators.

Prof. Dr. Karin Coninx is full professor at Hasselt University (UHasselt), Belgium. She obtained a PhD in sciences, computer science after a study of Human-Computer interaction (HCI) in immersive virtual environments. Her research interests include user-centred methodologies, persuasive applications in the context of eHealth, technology-supported rehabilitation, serious games, (multimodal) interaction in virtual environments, haptic feedback, intelligibility, mobile and context-sensitive systems, interactive work spaces, and the model-based realisation of user interfaces.
Karin Coninx has co-authored more than 300 international publications in scientific journals and conference proceedings. She teaches several courses on computer science and specific HCI subjects at Hasselt University, initiated a Master in HCI and co-initiated master profiles in Health Informatics and Engineering Interactive Systems. She presides the Interfaculty Board of the School for Educational Studies at Hasselt University (since 2017).

Prof. Dr. Frieda Steurs is full professor at the KU Leuven, Faculty of Arts, campus Antwerpen. She works in the field of terminology, language technology, specialized translation and multilingual document management. She is a member of the research group Quantitative Lexicology and Variation Linguistics (QLVL). Her research includes projects with industrial partners and public institutions. She is the founder and former president of NL-TERM, the Dutch terminology association for both the Netherlands and Flanders. She is also the head of the ISO TC/37 standardization committee for Flanders and the Netherlands.. She is the president of TermNet, the International Network for Terminology (Vienna). Since 2016, she is the head of research of the INT, the Dutch Language Institute in Leiden. In this capacity, she is responsible for the collection, development and hosting of all digital language resources for the Dutch Language. The INT is the CLARIN centre for Flanders, Belgium.

SCATE: A Smart Computer-Aided Translation Environment

We present SCATE: A Smart Computer-Aided Translation Environment developed in the SCATE research project. It is a carefully designed prototype of the user interface of a translation environment, that displays different translation suggestions coming from different resources in an intelligible and interactive way. Our environment contains carefully designed representations that show relevant context to clarify why certain suggestions are given. In addition, several relationships between the source text and the suggestions are made explicit, to help the user understand how a suggestion can be used and select the most appropriate one. Well designed interaction techniques are included that improve the efficiency of the user interface. The suggestions are generated through different web services, such as translation memory (TM) fuzzy matching, machine translation (MT) and support for terminology. A lookup mechanism highlights terms in the source segment that are available with their translation equivalents in the bilingual glossary.

Confirmed workshops (updated to 17 August)

David Filip

David Filip is Chair (Convener) of OASIS XLIFF OMOS TC; Secretary, Editor and Liaison Officer of OASIS XLIFF TC; a former Co-Chair and Editor for the W3C ITS 2.0 Recommendation; Advisory Editorial Board member for the Multilingual magazine; Co-Chair of the Standards Interest Group at JIAMCATT. His specialties include open standards and process metadata, workflow and meta-workflow automation. David works as a Moravia Fellow at the ADAPT Research Centre, Trinity College Dublin, Ireland. Before 2011, he oversaw key research and change projects for Moravia’s worldwide operations. David held research scholarships at universities in Vienna, Hamburg and Geneva, and graduated in 2004 from Brno University with a PhD in Analytic Philosophy. David also holds master’s degrees in Philosophy, Art History, Theory of Art and German Philology.

Does your tool support XLIFF 2?

This will be a hands on workshop for up to 15 participants. Participants will be required to bring their own laptops and supply their own legally obtained and valid licenses (can be evaluation licenses) of tools whose XLIFF 2 capability they’ll want to verify in the workshop. In the workshop, participants will learn about free and open source tools that can be used to verify XLIFF 2 support in CAT tools AND how to use those tools for that end. The resources employed will be the OASIS XLIFF TC hosted XLIFF 2 test suites, OKAPI XLIFF Toolkit Lynx validation service and command line tool. oXygen XLIFF framework, OASIS hosted XLIFF 2 validation artifacts, FREME Framework XLIFF services, Microsoft XLIFF 2 Object Model, Xmarker validation service etc. This workshop will provide actionable knowledge and skills with regards to validating XLIFF 2 support thus giving the participants the practical ability to verify tool makers’ marketing claims about their XLIFF 2 support.

Joss Moorkens and Kristiina Abdallah

Joss Moorkens is an Assistant Professor of Translation Studies in the School of Applied Language and Intercultural Studies in Dublin City University (DCU) and a researcher in the ADAPT Centre and the Centre for Translation and Textual Studies. Within ADAPT, he has contributed to the development of translation tools for both desktop and mobile. He is co-editor of a book on human and machine translation quality and evaluation (due in 2018) and has authored journal articles and book chapters on topics such as translation technology, post-editing of machine translation, human and automatic translation quality evaluation, and ethical issues in translation technology in relation to both machine learning and professional practice.

Kristiina Abdallah has worked as a translator, subtitler and technical writer. Since 2001 she has held various positions at the University of Tampere, namely that of an assistant, a lecturer and researcher. As of 2010 she has worked as a university teacher, first at the University of Eastern Finland, and currently at the Universities of Vaasa and Jyväskylä. She defended her doctoral thesis entitled Translators in Production Networks. Reflections on Agency, Quality and Ethics in 2012. Her research interests are translation sociology and, more specifically, translators’ workplace studies.

Freelance translation and the computer

The profession of translation has moved at a greater pace than many others towards contingent and freelance work. These work practices, coupled with greater use of technology, continue to change the nature of translation work, with associated effects on translator’s agency and autonomy. Following on from our proposed main-track session on agency and work practices in translation production networks, in this workshop we will present a short summary of our work and thoughts on contingent work in translation, highlighting opportunities to maximise agency. The main purpose of this session, however, is to facilitate a discussion about working standards for translators and satisfaction with technology, and the effects of these influences on job satisfaction and translation quality. Related to these are considerations of productivity expectations and income, and whether participants’ experiences with both are realistic and fair. Finally, we wish to consider how translators’s agency and sense of purpose could be maximised in translation production networks.

Gloria Corpas Pastor

Gloria Corpas Pastor, BA in German Philology (English) from the University of Malaga. PhD in English Philology from the Universidad Complutense de Madrid (1994).

Visiting Professor in Translation Technology at the Research Institute in Information and Language Processing (RIILP) of the University of Wolverhampton, UK (since 2007), and Professor in Translation and Interpreting (2008). Published and cited extensively, member of several international and national editorial and scientific committees. Spanish delegate for AEN/CTN 174 and CEN/BTTF 138, actively involved in the development of the UNE-EN 15038:2006 and currently involved in the future ISO Standard (ISO TC37/SC2-WG6 “Translation and Interpreting”).

Regular evaluator of University programmes and curriculum design for the Spanish Agency for Quality Assessment and Accreditation (ANECA) and various research funding bodies.

President of AIETI (Iberian Association of Translation and Interpreting Studies), member of the Advisory council of EUROPHRAS (European Society of Phraseology) and Vice-President of AMIT-A (Association of Women in Science and Technology of Andalusia).

VIP: Voice-text integrated system for interpreters

Interpreting is an extremely strenuous task, since they must devote much effort in terms of decoding, memorising and encoding a message. Interpreters should, as translators and other language professionals do, benefit from the development of technology and, thereby, enjoy considerable improvement of their working conditions. However, currently their work relies by and large on traditional or manual methods, and the technological advances in interpreting have been extremely slow.

Unlike translators, for whom a myriad of computer-assisted tools are available, interpreters have not benefited from the same level of automation or innovation. Fortunately, there is a growing interest in developing tools addressed at interpreters as end users, although the number of these technology tools is still very low and they are not intended to cover all interpreters needs.

The goal of the VIP project is to revolutionise the interpreting industry by creating an interpreting workbench tool which will have the same effect that language technologies for translators have had in the translation industry in recent decades. To this end, we intend to (a) identify the
real needs of interpreters and how and to what extent their work can be automated, (b) survey existing interpreting technologies, and (c) develop the first integrated system to enhance the productivity of the work of interpreters (professional, trainers and trainees), both during the
interpretation process and in the preparation of various interpretation tasks.

Anja Rütten

Dr Anja Rütten (Sprachmanagement.net) is a freelance conference interpreter for German A, Spanish B, English and French C based in Düsseldorf, Germany since 2001. Apart from the private market, she works for the EU institutions and is a lecturer at the TH Cologne. She holds a degree in conference interpreting as well as a phD of the University of Saarbrücken (doctoral thesis on Information and Knowledge Management in Conference Interpreting, 2006). As a member of AIIC, the international conference interpreters’ association, she is actively involved in the German region’s working group on profitability. She has specialised in knowledge management since the mid-1990s and shares her insights in her blog on www.dolmetscher-wissen-alles.de.

Terminology management tools for conference interpreters – Current tools and how they address the specific needs of interpreters

Ever since the 1990s, sophisticated terminology management systems have offered a plethora of data fields and management functions to translators, terminologists, and conference interpreters. Nevertheless, interpreter-specific tools have been developed in parallel to suit their special needs. They were mostly inspired, at least initially, by one or very few users and developed by a single developer or a very small team.

The intention of this workshop is to give interpreters and other interested participants an overview on which tools are available for their terminology work, highlighting the pros and cons of each of them. Providers and developers of terminology management systems will get valuable insight on the specific needs of conference interpreters and the reasons why, if using terminology management systems at all, conference interpreters tend not to use the sophisticated term databases translators or terminologists use.
Some of the tools presented are not being further developed any more, and no support is offered. As they still run perfectly well in the versions at hand, they will nevertheless be shown in order to complete the picture. Due to time restrictions, some tools will be shown “live” and others with the help of screenshots.

Due to time restrictions, only the most relevant aspects of terminology management in conference interpreting will be addressed. Which solution is best for filtering and categorising my terminology? Which one offers the best search function for the booth? Which one is best for sharing glossaries and online collaboration, or most convenient for mobile use? Information on price models and supported operating systems will also be provided.
The programs presented will be

  • Glossarmanager by Glossarmanager GbR/Frank Brempe, Bonn, Germany
  • Glossary Assistant by Reg Martin, Switzerland
  • Interplex by Peter Sand, Eric Hartner, Geneva, Switzerland
  • InterpretBank by Claudio Fantinuoli, Germersheim, Germany
  • Interpreters’ Help by Benoît Werner/Yann Plancqueel, Berlin/Paris, Germany/France
  • Intragloss by Dan Kenig and Daniel Pohoryles, Paris, France
  • Lookup by Christoph Stoll, Heidelberg, Germany
  • Terminus by Nils Wintringham, Zürich, Switzerland

If time allows, generic solutions like Microsoft Excel/Access and Google Sheets will also be discussed as an alternative to interpreter-specific tools.

Andrzej Zydroń

Andrzej Zydroń MBCS CITP

CTO @ XTM International, Andrzej Zydroń is one of the leading IT experts on Localization and related Open Standards. Zydroń sits/has sat on, the following Open Standard Technical Committees:

1. LISA OSCAR GMX
2. LISA OSCAR xml:tm
3. LISA OSCAR TBX
4. W3C ITS
5. OASIS XLIFF
6. OASIS Translation Web Services
7. OASIS DITA Translation
8. OASIS OAXAL
9. ETSI LIS
10. DITA Localization
11. Interoperability Now!
12. Linport

Zydroń has been responsible for the architecture of the essential word and character count GMX-V (Global Information Management Metrics eXchange) standard, as well as the revolutionary xml:tm (XML based text memory) standard which will change the way in which we view and use translation memory. Zydroń is also chair of the OASIS OAXAL (Open Architecture for XML Authoring and Localization) reference architecture technical committee which provides an automated environment for authoring and localization based on Open Standards.
Zydroń has worked in IT since 1976 and has been responsible for major successful projects at Xerox, SDL, Oxford University Press, Ford of Europe, DocZone and Lingo24 in the fields of document imaging, dictionary systems and localization. Zydroń is currently working on new advances in localization technology based on XML and linguistic methodology.
Highlights of his career include:
1. The design and architecture of the European Patent Office patent data capture system for Xerox Business Services.
2. Writing a system for the automated optimal typographical formatting of generically encoded tables (1989).
3. The design and architecture of the Xerox Language Services XTM translation memory system.
4. Writing the XML and SGML filters for SDL International’s SDLX Translation Suite.
5. Assisting the Oxford University Press, the British Council and Oxford University in work on the New Dictionary of the National Biography.
6. Design and architecture of Ford’s revolutionary CMS Localization system and workflow.
7. Technical Architect of XTM International’s revolutionary Cloud based CAT and translation workflow system: XTM.

Specific areas of specialization:
1. Advanced automated localization workflow
2. Author memory
3. Controlled authoring
4. Advanced Translation memory systems
5. Terminology extraction
6. Terminology Management
7. Translation Related Web Services
8. XML based systems
9. Web 2.0 Translation related technology

The Localization Industry Word Count Standard: GMX-V

Word and character counts are the basis of virtually all metrics relating to costs in the L10N Industry. An enduring problem with these metrics has been the lack of consistency between various computer assisted tools (CAT) and translation management systems (TMS). Notwithstanding these inconsistencies there are also issues with common word counts generated by word processing systems such as Microsoft Word. Not only do different CAT and TMS systems generate differing word and character counts, but there is also a complete lack of transparency as to how these counts are arrived at: specifications aren’t published and systems can produce quite widely different metrics. To add clarity, consistency and transparency to the issue of word and character counts the Global Information Management Metrics Volume (GMX-V) standard was created. Starting with version 1.0 and then as version 2.0 GMX-V addresses the problem of counting words and characters in a localization task, and how to exchange such data electronically. This workshop goes through the details of how to identify and count words and characters using a standard canonical form, including documents in Chinese, Japanese and Thai, as well as how to exchange such data between systems.

Anja Rütten, Alexander Drechsel and Josh Goldsmith

Danielle D’Hayer is an associate professor in interpreting studies at London Metropolitan University. She is the course director of the MA Conference Interpreting, the MA Interpreting, MA Public Service Interpreting and interpreting short courses that include a Training the Trainers for Interpreting Studies programme and a portfolio of Continuous Professional Development (CPD) activities. These courses, which she developed single-handedly, have attracted both professional interpreters and novices from the UK and abroad.
Danielle researches communities of practice for interpreting studies. Her main interests include innovative ways to enhance formal and informal blended leaning using social media, new technologies and on-line platforms. You can follow her on Twitter @DDhayer.

Dr Anja Rütten (Sprachmanagement.net) is a freelance conference interpreter for German A, Spanish B, English and French C based in Düsseldorf, Germany since 2001. Apart from the private market, she works for the EU institutions and is a lecturer at the TH Cologne. She holds a degree in conference interpreting as well as a PhD of the University of Saarbrücken (doctoral thesis on Information and Knowledge Management in Conference Interpreting, 2006). As a member of AIIC, the international conference interpreters’ association, she is actively involved in the German region’s working group on profitability. She has specialised in knowledge management since the mid-1990s and shares her insights in her blog on www.dolmetscher-wissen-alles.de .

Alexander Drechsel has been a staff interpreter with the European Commission’s Directorate-General for Interpretation since 2007. He has studied at universities in Germany, Romania and Russia and his working languages are German (A), English (B), French and Romanian (C). Alexander is also a bit of a ‘technology geek’ with a special interest in tablets and other mobile devices, and regularly shares his passion and knowledge with fellow interpreters during training sessions and on the web at http://www.tabletinterpreter.eu/ .

Joshua Goldsmith is an EU-accredited interpreter working from Spanish, French, Italian and Catalan into English. He splits his time between interpreting and working as a trainer and researcher at the University of Geneva, where he focuses on the intersection between interpreting, technology and education. A lover of all things tech, Josh shares tips about technology and interpreting in conferences and workshops, the Interpreter’s Toolkit column ( https://aiic.net/search/tags/the-interpreter’s-toolkit ), and on Twitter (@Goldsmith_Josh).

Marcin Feder has been an interpreter at the European Parliament since 2003 and the Head of the Polish Interpretation Unit from 2012 to 2016. He is now the Head of Interpreter Support and Training Unit and the acting Head of Multilingualism and Succession Planning Unit. He studied at Adam Mickiewicz University in Poznan, Poland (MA in English and PhD in Linguistics focusing on Computer Assisted Translation) and Monterey Institute of International Studies, USA (Junior Fulbright Scholarship). These days, apart from the regular managerial duties, his main interests are the use of tablets in the booth, new technologies to support interpreters in their daily work and all things paper-smart. He is also an avid runner.

Barry Slaughter Olsen is a veteran conference interpreter and technophile with over two decades of experience interpreting, training interpreters and organizing language services. He is an associate professor at the Middlebury Institute of International Studies at Monterey (MIIS), the founder and co-president of InterpretAmerica, and General Manager of Multilingual Operations at ZipDX. He is a member of the International Association of Conference Interpreters (AIIC). Barry is the author of “The Tech-Savvy Interpreter” a monthly column and video series published in Jost Zetzsche’s Tool Box Journal focusing on interpreting technology. For updates on interpreting, technology and training, follow him on Twitter @ProfessorOlsen.

PANEL DISCUSSION : New frontiers in interpreting technology

Technology has the potential to shape the future of interpreting. Indeed, it has already begun to do so. From tools that assist interpreters to devices that may replace them altogether, technologies for delivering interpreting services to tools to teach interpreting, this panel discussion will span the gamut of technology in interpreting, considering current developments and future innovations.
A panel of leading practitioners, researchers and trainers with experience in the private and institutional markets will invite audience members to engage with the state-of-the-art of technology in our industry.
Topics will include:
● enhancements to conference technology, including digital consoles, bone-conduction headphone, and efforts to leverage mobile phone technology for simultaneous interpreting;
● advances in collaborative glossary management;
● the impact of remote interpreting delivery platforms on different market segments;
● how software, apps and mobile device can be used in all phases of the interpreting process;
● tools for online and technology-assisted interpreter training, including SmartPens and online communities like Speechpool and InterpretimeBank; and
● innovations in virtual reality.
Join us for this interactive conversation about the present and future of technology for interpreters, and consider how these technologies may shape your personal practice and the industry as a whole.