Presenters and Co-authors

(by alphabetical order of first presenter)

Name

Title

When

Argelia Peña Aguilar (Univ. of Ottawa)

Argelia Peña Aguilar has been an Associate professor at the University of Quintana Roo (Mexico) for eleven years.  She has taught English Language and Translation/Interpreting courses from English into Spanish in the Language and Education Department.  She is currently studying for a PhD in Translation Studies at the University of Ottawa (second year), and her research interests revolve around translation technologies training, and feminist translation studies.

Poster: Usefulness of Translation Technology Training from Mexican Universities

In a study done in 2018 it was reported that few professors teach technology in few translation courses in Mexico. Some reasons for this were that instructors had not been well trained in their academic programs when they were students, or they lacked a more comprehensive knowledge of these technologies (Peña, 2018). Effective training was not possible for most of these instructors as students and they seem to be reproducing similar learning insufficiencies with future translators. Because of this, another survey-based project was devised to identify the use that professionals who graduated from Mexican translation programs are making of translation technologies.

How has their educational background affected their disposition towards the use of translation technologies? Some results indicate that professional translators do not resort to the use of “core” translation technologies very often, but do use other electronic resources useful for accomplishing their tasks. One in two translators thinks their income has increased due to their technology knowledge, and they learn about these technologies on their own. Professional translators think they could have learned about Translation Environment Tools ((TEnTs) at university (and they wished they had), but university instructors are still not teaching these technologies as much. So there is a need reported by a few professionals, but not being dealt by some university programs.

Keywords: translation technologies, translation training, TEnTs, translation environment tools, computer-aided translation, CAT, Mexico

Peña, A. (2018). Use of technologies in Mexican translation programs. Unpublished manuscript.

Friday 22 at 14.15
Education Room
Josep Bonet (WTO)

Josep Bonet was not meant to be a translator, but rather a chemist. Something went wrong, though, and finished by spending almost 30 years in the Directorate-General for Translation of the European Commission. He played most the roles available, translator, help desk officer, information officer, communication officer, manager of units concerned with translation, IT, language technologies and terminology, learning & development, knowledge management, etc. He chaired the JIAMCATT forum, where international organisations discuss language technology. Presently he is Director of the languages, Documentation and Information Management Division of the World Trade Organization. He pretends that abandoning active translation allowed the average quality of translation output to increase substantially.

Innovation in the International Organisation: Can we Do Better?

The modern economy is all about innovation, disruption, shifting paradigms, accelerating the pace of change. International Organizations may be perceived as not following this trend. Is this true? How do they react to changes in their environment? This presentation will give some insights, based on experience in two such organizations, one very large and another of medium size.

Friday 22 at 14.15
Lecture Theatre

Berrnardette Castaneira Legrand (Univ. of Ottawa)

Certified translator for the State of Baja California, Mexico; Associate Professor in Translation Studies at the Faculty of Languages, Universidad Autonoma de Baja California (UABC) Mexico; second year PhD student in Translation Studies, University of Ottawa, Canada. Research fields of interest include: Legal dimensions of translation, Translation Technologies, and Terminology.

Poster: Legal Translators and the Use of Translation Technologies

The present study examines the use of language technologies in legal translation, according to the findings of existing results and in designing and implementing a new study in northern Mexico. Considering that the use of technology has become a fundamental skill for professional translators, it is essential to bridge the gap between legal translators and language technologies. Legal translators’ uptake of technology has not always been positive. This research was conducted online with participants from Baja California, and participants are certified legal translators for the State; the participants’ main working languages are Spanish and English. The participants answered a questionnaire, which helped to identify the translators’ profiles, their use of electronic resources and adoption or non-adoption of language technologies. Data gathered was analysed in a qualitative manner. Main findings have shown that some legal translators implement translation technologies in their work. We have been able to identify, to some extent, the tools that legal translators are using. Although the majority of translators reported using electronic dictionaries, term banks, web tools, and search engines, some translators have already started to implement to CAT tools, such as terminology management systems, translation memory systems, term extractors, active terminology recognition tools, concordancers, bitexts and aligners, and localization tools.

Friday 22 at 14.15
Education Room
Caroline Champsaur (OECD)

Caroline Champsaur has been working at the OECD for almost 20 years as the Head of the Reference and Terminology Unit (Translation Division). Over the years, she led the change from paper to digital. As a Counsellor for Digital, she also manages projects on Terminology and Machine Translation and participates to several projects of OECD’s Digital Strategy.

She holds a PhD in computational linguistics and a Master’s Degree in computer science (University of Paris 7, France), as well as a Master’s Degree in German Language and Literature (University of Paris 7, France). She also studied Artificial Intelligence (Aachen University of Technology, Germany), German Language (Westfälische Wilhelms-Universität Münster, Germany) and French as a Foreign Language (University of Paris 7, France).

 

Workshop: OECD Neural Machine Translation Pilot Project : Methodology and Results

The OECD (Organisation for Economic Co-operation and Development) has launched a Digital Strategy in order to maximize the business benefits of digital initiatives. In 2018, one of the Translation Division’s contributions to this program was the development of a Neural Machine Translation (NMT) system in collaboration with the WIPO (World Intellectual Property Organization). The objectives of this Pilot project was to explore new sources of efficiency gains and to extend the coverage of OECD content in both official languages (English and French).

This workshop will explain what has been achieved with this project. It will describe the methodology chosen to assess the quality of the sentences translated automatically and share the results and the lessons learned. Hopefully, it should help the participants decide how to take advantage of this new technology.

Friday 22 at 10.45
Education Room

Denis Dechandon (EPO), Maria Recort Ruiz (ILO) and Aniko Gerencser (EPO)

Denis Dechandon has over 20 years’ experience in translation and linguistics, in office automation and in different management roles. After getting acquainted with the translation work and its requirements at EU level, he fully committed himself to the definition and implementation of processes and workflows to provide structured and efficient support to linguists and to streamline the work of support teams.

Previously in his last role, Denis was responsible for leading a service dedicated to the linguistic and technical support provided to translators, revisers, editors, captioners and  subtitlers (Computer Assisted Translation, corpus management, formatting and layouting, machine translation and terminology). Additionally, he supervised the maintenance and development of tools and linguistic resources at the Translation Centre for the Bodies of the European Union. Committed to further changes and evolutions in these fields, Denis took over the role of InterActive Terminology for Europe (IATE) Tool Manager from May 2015 to August 2017.

In his current role as Head of the Metadata sector of the Publications Office of the European Union, he is leading the activities in standardization (in particular: EuroVoc and registry of metadata) as well as intensely involved in the field of linked open data at the Publications Office of the European Union. Latest projects involve the development of synergies between several different stakeholders, such as EU institutions, agencies and bodies, international organisations and national public services.

Maria Recort Ruiz is a philologist, translator and terminologist who works as Document Services Coordinator and Terminology Manager at the International Labour Organization in Geneva. She is responsible for the production and management of official documents, management of terminology work and the use of new CATT tools to improve working methods. She holds a Degree in Slavic Philology from the University of Barcelona, where she specialized in Russian and Polish Language and Literature, and Linguistics; a Master in French and Comparative Literature (19th-20th centuries) from the University of Montpellier, where she conducted research on the roman populaire at the beginning of the 20th century; and a Master in Specialized Translation from the University of Geneva. Before joining the ILO, she worked as a freelance translator and editor for international organizations and the private sector.

Anikó Gerencsér holds a Master`s Degree in Library and Information Science and a PhD in Italian Language and Literature from the University ELTE of Budapest.

Since joining the Publications Office of the European Union she has been working in the field of metadata standardisation and linked open data management. Her particular area of responsibility is the co-maintenance of the EuroVoc multilingual thesaurus and its alignment with other controlled vocabularies. She currently works on the optimisation of the thesaurus management tool Vocbench3 which involves analysing users` needs and improving collaborative features. She is in charge of providing presentations, consultancy and trainings regarding the use of VocBench3 for EU institutions. In addition she strongly contributes to an on-going project that aims to achieve interoperability between controlled vocabularies by sharing common tools and formats for the creation, use and maintenance of vocabularies and taxonomies.

Terminology: Towards a Systematic Integration of Semantics and Metadata
Friday 22 at 14.15
Lecture Theatre
Ana-Luz Diaz and Simone Maier (Univ. of Applied Sciences Würzburg-Schweinfurt)

Ana-Luz Diaz M.A. in Translation and Interpreting (ES-EN-DE) from the University of Granada, M.A. in Journalism, from the University CEU Madrid.

Lecturer in Specialised Translation (DE>ES), Intercultural Communication and CAT Tools at FHWS University in Germany.

Simone Maier, B.A. in Translation Studies from Heidelberg University, M.A. in Specialised Translation (DE-EN-ES), worked for several years as project manager and in-house translator.
Lecturer in Specialised Translation (EN>DE) for Software Localization and CAT Tools at FHWS University in Germany.

Poster: Machine Translation vs. Human Translation: An Analysis of the Use and Impact of Pre-Editing in a Variety of Text Types

Post-editing has become a regular part of the machine translation process, and many translators are already specialised in providing this service. Post-editing is, however, not the only way to improve the outcome of machine translation. An alternative is the pre-editing of a text prior to the utilisation of machine translation software. The poster describes the results of an independent study project in a course for translation students working with the language pairs English-German and English-Spanish. The task was to analyse specific problems of MT with regard to text type and terminology, with a view to discovering how pre-editing a text can improve the result of MT. The participants first produced a human translation of a text they had chosen themselves. They then carried out a translation using DeepL and assessed the problems of the resulting texts. In order to reduce mistakes and to achieve a better machine-translated version, they pre-edited the source texts and compared the revised outcome to the previous versions.

Overall, it can be said that in many cases the translation by DeepL was surprisingly good and could be further improved by pre-editing. Some problems could easily be solved by pre-editing. This was achieved in particular by replacing terms that could have several meanings by terms which only imply the desired meaning, and that sentences in which the references were not clear reformulated so that the references were clear. This was in line with the general rules for pre-editing, that ambiguous terms should be avoided, references should be clearly identifiable, and complex sentences should be simplified. The translation results that were achieved by pre-editing the texts were surprisingly good in some cases, but there are certain problems that cannot be solved without post-editing. The inconsistent use of terms, for example, is an almost unmanageable problem. A human translator is therefore still essential for a good translation.

Thursday 21 at 15.15
Education Room
Emmanuelle Esperança-Rodier and Caroline Rossi (Univ. of Grenoble-Alpes)

Emmanuelle Esperança-Rodier is a lecturer at Univ. Grenoble Alpes (UGA), France, where she teaches English for Specific Purposes, and a member of the Laboratoire d’Informatique de Grenoble (LIG). After defending a PhD in computational linguistics, on “Création d’un Diagnostique Générique de Langues Contrôlées, avec application particulière à l’Anglais Simplifié”, she worked as a post-editor in a translation agency. Back at University, she participated in IWSLT and WMT evaluation campaigns, as well as in several LIG projects. She now works on the evaluation of MT systems based on competences and focused on tasks, translation error analysis and multilinguism.

Time is everything: A comparative study of human evaluation of SMT vs. NMT

Translation process research has developed tools to gather and analyse empirical data, but while a variety of measures have proved useful and reliable to measure post-edit machine translation effort (see e.g. Vieira 2016 : 42), translation processes are seldom considered when assessing the relevance of a given Machine translation post-editing scenario. Our study seeks to determine the impact of including MTPE in the evaluation process. We selected adequacy and fluency ratings. Based on two distinct experimental conditions, we then compared the ratings produced without performing PE and those produced immediately after a light PE process. Inter-rater reliability was assessed for each segment in each text (N=55) using Fleiss’ kappa for adequacy and fluency scores, and an intraclass correlation coefficient (Vieira 2016 : 52) for temporal measures. While the reliability of the measures collected without PE was low, the measures collected in PET were for the most part homogeneous. Qualitative analyses of the problematic segments, as evidenced by both kappa and intraclass correlation coefficients, showed strong Spearman’s correlations, whether positive or negative, between temporal measures and all the other metrics for NMT but weakest ones for SMT. Based on these results, we discuss the advantages and risks of NMTPE.


Accolé: A Collaborative Platform of Error Annotation for Aligned Corpora

This article presents a platform, named ACCOLÉ, for the collaborative annotation of translation errors.

ACCOLÉ offers a range of services that allow simplified management of corpora and typologies of errors, annotation of effective errors, collaboration during annotation, and finally different kinds of search in corpora. ACCOLÉ allows the annotation of translation errors according to built-in error typologies, Vilar’s typology or DQF-MQM or uploaded ones, on several corpora of different texts, translated by different Statistical or Neural Machine Translation systems, as well as processing the annotated corpora created in order to look for typical error models and patterns, related to a specific MT system.

The collaboration feature also gives the possibility to detect any misleading interpretation of an error type among the annotators. ACCOLÉ currently provides 15 corpora, 7 projects of 201,474 words and 18,301 annotations that we will describe in the final paper. Eventually, we will implement the semi-automatic propagation of found patterns on other corpora to enlarge the scope of linguistic studies, thus providing to the community a wide range of error annotated bilingual parallel corpora.

Friday 22 at 10.45
Lecture Theatre

Friday 22 at 14.15
Lecture Theatre

Sabrina Girletti, Pierrette Bouillon, Martina Bellodi and Philipp Ursprung (Univ. of Geneva)

Sabrina Girletti is a PhD student at the Translation Technology Department of the Faculty of Translation and Interpreting (FTI) at the University of Geneva, where she contributes to postgraduate courses in machine translation and post-editing. Her research interests include post-editing approaches and human factors in machine translation. As a young language technology consultant, she also collaborates with Suissetra, the Swiss association for translation technology promotion. She is currently involved in projects testing the implementation of machine translation at several corporate language service departments in Switzerland. Sabrina holds a master’s degree in Translation with a specialisation in Translation Technology from the University of Geneva and a bachelor’s degree in Linguistic and Cultural Mediation from the University of Naples L’Orientale.

Pierrette Bouillon has been Professor at the FTI, University of Geneva, since 2007. She is currently Director of the Department of Translation Technology (referred to by its French acronym TIM) and Dean of the Faculty. She has numerous publications in computational linguistics and natural language processing, particularly within lexical semantics (Generative lexicon theory), speech-to-speech machine translation for limited domains and, more recently, pre-editing and post-editing. Between 2012 and 2015, she coordinated the European ACCEPT project (Automated Community Content Editing PorTal). At present, she co-coordinates the new Swiss Research Center for Barrier-free communication with the Zurich University of Applied Sciences, and the project BabelDr with the HUG (Geneva University Hospitals). She also takes part in the new COST network EnetCollect: European Network for Combining Language Learning with Crowdsourcing Techniques.

Martina Bellodi graduated from the University of Bologna in 2003 and began her career as a freelance translator. In 2009 she started working as an in-house translator at Swiss Post Language Services. She was promoted to Head Translator in 2011 and to Deputy Head of Language Services in 2012. Since 2014 Martina has been in charge of Language Services’ operational and strategic management. She holds an EMBA degree from the University of Applied Sciences and Arts of Southern Switzerland (SUPSI) and has appeared as a keynote speaker at several industry conferences (tcworld Stuttgart, LQA Symposium Zurich, XTM Live Amsterdam).

Philipp Ursprung holds a degree in Translation and translation technology from the University of Surrey (UK). Before joining Swiss Post as language technology specialist in 2018, he has worked in the localization industry for more than 10 years in different positions with both language services providers and corporate language services departments, where he focused on project management and the introduction of TMS and MT systems and optimization of translation processes and workflows.

Preferences of End-Users for Raw and Post-Edited NMT in a Business Environment

The in-house Language Service at Swiss Post translates a wide variety of texts from and into German, French, Italian, and English. It has emerged from internal discussions over the years that the extensive hype around neural machine translation (NMT) and its improved fluency, in comparison with previous approaches, has led many of Swiss Post’s employees to turn to freely available, generic MT systems to obtain quick, raw translations.

Before introducing NMT into their production workflow, Swiss Post’s Language Service decided to carry out a study to assess whether their customers (Swiss Post employees, who are also end-users of translations produced by the Language Service) would rate post-edited NMT more highly than raw NMT for both a customized and a generic MT engine (DeepL). Most importantly, the study also assessed whether the customers would be willing to pay for post-edited texts when made aware of some production metadata, such as data security and cost. This latter aspect could help determine whether the customers would still value the human intervention or whether they would rather accept a lower quality translation and associated risks if it means they can save on costs.

Friday 22 at 11.45
Lecture Theatre
Christopher Gledhill and Maria Zimina (Univ. Paris-Diderot)

Christopher Gledhill is Professor of English Linguistics at the Université Paris Diderot (Paris, France) where he is currently director of Languages for Specialists of all Disciplines (LANSAD), coordinator of Masters in Applied Foreign Languages (Mention LEA) and cocoordinator with Natalie Kübler of a research Masters in Languages for Specific Purposes, Corpus Linguistics and Translation studies (Master LSCT). He currently teaches and conducts research in interlinguistics, specialised translation, phraseology, systemic functional grammar and text linguistics.

Maria Zimina-Poirot is Associate Professor in English Language Studies at Paris Diderot (Université Paris Diderot). She holds a PhD in Language Studies and Linguistics from Paris 3 – Sorbonne nouvelle University (2004). Her PhD thesis focused on developing new tools for textometric exploration of multilingual text corpora. Between 2005-20012 she worked as a teaching and research assistant at Paris Nord – Paris 13 University, INaLCO, INSERM, etc. and as a technical writer for Orange Business Services (Paris). Her current teaching and research activities concern textometric analysis of multilingual corpora, text typologies, controlled languages, terminology, technical writing and computer-aided translation.

The Impact of Machine Translation on a Masters Course in Web Translation: From Disrupted Practice to a Qualitative Translation/Revision Workflow

The introduction of technology into translation curricula is a complex task in terms of translation competences and their acquisition. Computer tools and MT directly affect trainee translators.

This study investigates the impact of technology on students on a Master’s in Specialised Translation and Language Industries at Université Paris Diderot. We present the results of a teaching project “Website translation into English” which places strong emphasis on hands-on applications of MT.

The aim of the project is to provide students with a semi-professional work experience in which they face real-life website translation problems. Students are expected to translate and revise webpages from French into English using a professional platform SystranLINKS. The first results of our study show that a more equipped translator’s workstation results in assisted but also disrupted translation practice, and requires additional learning/teaching time. Intensive practice of MT raises students’ awareness of the importance of a revision workflow, and gives students a broader understanding of translation quality.

Our methodology involves the analysis of project reporting forms, which students write at the end of the course as a record of their learning experience. We examine both their explicit comments and their implicit metalanguage, in order to explore how they conceptualise MT.

Friday 22 at 16.45
Lecture Theatre
Lucía Guerrero and Kirill Soloviev (CPSL and ContentQuo)

Lucía Guerrero is a Machine Translation Specialist at CPSL, a linguistic services provider based in Spain with presence in Germany, the UK and the US. The range of services includes translation, software and web localization, multilingual SEO, interpreting, multimedia and e-learning in all major Western and Eastern European, Scandinavian, Asian and Middle-Eastern languages. Lucía is also part of the collaborative teaching staff at the Universitat Oberta de Catalunya. Having worked in the translation industry since 1998, she has also been a senior Translation and Localization Project Manager specialized in international institutions, has managed localization projects for Apple Computer and has translated children’s and art books.

Kirill Soloviev is the Co-Founder & CEO at ContentQuo, an Estonian tech startup helping Global Top-10 LSPs, enterprise loc teams, and government agencies reduce translation quality risk, improve vendor performance, and boost MT quality at any scale, regardless of their TMS. During his 16-year industry career, Kirill served in diverse buyer-side & vendor-side roles, most recently as Global Director of Localization at Acronis, a $150M data protection and disaster recovery software company. Kirill also co-organises Localization Unconference in Tallinn, collaborates with TAUS, and loves consulting both new and seasoned localization pros about their careers.

Machine Translation Evaluation at CPSL with ContentQuo

Evaluating the performance of an MT system with new content – that is, MT performance prediction – is one of the most challenging aspects of MT, mainly because lack of reference translations does not allow using automatic metrics. Additionally, human evaluation can be expensive and time-consuming. As an LSP, at CPSL we deal with hundreds of translation requests daily and must choose the most appropriate workflow for our customers in a timely manner. That’s why we needed a fast, reliable and cost-effective solution allowing us to find out if a given MT system is suitable for specific content. After trying different methods and tools, we chose the solution provided by ContentQuo, a translation quality management platform, based on the widely accepted Adequacy-Fluency methodology for MT evaluation. In our presentation we will introduce you to the challenges of MT quality evaluation and how we address them with ContentQuo.

Friday 22 at 9.15
Education Room
Jochen Hummel (Coreon)

Jochen Hummel is co-founder and CEO of Coreon, the leading SaaS solution for multilingual knowledge systems. He is CEO of ESTeam AB, a provider of language technology and semantic solutions to EU organisations and corporations. He serves as vice-chairman of LT-Innovate, the Forum for Europe’s Language Technology Industry. He has a software development background and had grown his first company, TRADOS, to the world leader in translation memory and terminology software. In 2006 he founded Metaversum, the inventor of the virtual online world Twinity and was its CEO until 2010. He is a well-known, internationally respected software executive and serial entrepreneur. He serves on boards and is mentor/angel for several start-ups in Berlin.


Keynote speech Friday 22 at 9.45
Lecture Theatre
Marion Kaczmarek and Michael Filhol (LIMSI)

Marion Kaczmarek is a PhD student in both linguistics and computer sciences, in the quite particular field of Sign Language. Former French Sign Language Interpreter, trained at the University of Rouen (France) where she also graduated with a Language Sciences Master’s degree, her studies have led her from cognitive sciences to Sign Language linguistics. Her PhD involves both Sign Language Translation and CAT software, trying to find ways of equipping the Sign Language translators with computer assistance. Her work started in 2018, joining the CNRS (the French nation scientific research center) and other associates on a more global project concerning media accessibility.

Michael Filhol is a computer scientist who has always been passionate about languages and linguistics. He naturally turned to NLP in his studies, which took place in France and Ireland after the year 2000, always learning new languages and comparing how they work as a side hobby.

He chose to focus on Sign Language for his PhD, the most exotic linguistic system he knew, and which he had been learning since high school. Largely still unknown to science and virtually absent in the NLP field, he addressed and proposed a formal description model of signs for Sign Language synthesis by 3D avatars. He defended his PhD in 2008 at Université Paris Sud (Orsay, France), and continued his research career on Sign Language processing.

He stayed for a post-doctoral year at Gallaudet University (Washington, DC, USA), where all classes and services are accessible in Sign Language, and some of the most famous researchers on Sign linguistics are hosted. Back in France, he got his permanent researcher position at CNRS (the French national scientific research centre), where he kept working on the formal description and computer implementation of Sign Language.

He eventually proposed AZee, a formal approach capturing all levels of discourse and capable of driving a 3D avatar to animate Sign from a combination of semantic operations. It is now used by the world leaders in Sign synthesis, as input for their animation platforms. While always improving and extending the coverage of the AZee approach, his research interests have grown to encompass more topics like graphical writing systems for Sign Language, or automatic and assisted text-to-Sign translation.

Assisting Sign Language Translation: What Interface Given the Lack of Written Form and the Spatial Grammar?

Computer-assisted translation (CAT) software offers tools for the translators to ease their tasks, and gain time as well as comfort. However, despite the growing need for Sign Language content, there has been no effort to equip Sign Language translation with CAT software. The problem we address here is the specification of such software. Sign Languages are visual and iconic, with grammar and discourse organisation, but also no written form. This is problematic when it comes to CAT, for it relies on editable written structures and the fact that the concatenation of the translated segments will result in the translation of the concatenated source segments (we call it the linearity assumption).

In this paper, we explain that Sign Language cannot follow those rules. We address those differences by means of new adapted modules which would be more flexible, and by considering new tools based on professionals’ feedback towards their actual practice as well as the problems they encounter during the translation process. We will detail those results along with the presentation of how we envisage a sign language concordancer, and its database.

]/tc41spoiler]
Thursday 21 at 15.15
Education Room
Rodolfo Maslias (European Parliament)
New Audiences for EU Terminology

A short PowerPoint presentation followed by live surfing in the public websites termcoord.euYourTerm.org and in the (password protected) EU interinstitutional terminology portal EurTerm, focusing on the terminology management in the European Parliament, the cooperation between EU Institutions at central (EurTerm) and at language levels (wikis), the collaboration with the interpreters in the EU for terminology, the interoperability of the new version of IATE, the efforts of TermCoord for a new terminologist profile in the recruiting procedures in the EU Institutions, the connection of the EU and other terminology resources (like the EP’s GlossaryLinks) with the NMT and the post editing and quality control software, terminology projects with Universities with and for IATE, Master courses on terminology at the Universities of Luxembourg (36 hours), Savoie-Mont Blanc (21 hours) and Orientale Napoli (11 hours)  and occasionally in many Universities (Vigo, Germersheim, ISIT Paris a.o.) and the new approach of “plain terminology” projects, adapted to communication needs and addressed to the civil society with the programme “Terminology without Borders” in several fields and in collaboration with specialised EU Agencies and International Organisations and with specialised departments of Universities in several European countries.

Lecture Theatre
Thursday 21 at 14.15
Jean-François Richard (Terminotix)

Having a background in automated productivity, Jean-François Richard has worked in the field of computer aided translation tools for the last thirty years. Through different experiences of work, he acquires expertise in translation memory systems, terminology extraction tools, terminology management, machine translation, full text, bitexts and project management systems. During his career, he develops SynchroTerm, a powerful terminology extraction tool allowing the feeding of terminology databases from translated documents. In September 2006, Jean-François Richard joined the Terminotix team where he occupied the Sales Director position. In April 2010, he acquired Terminotix and since then, he is the president of the company. Under his leadership, the company has grown at a rate of more than 20% per year.

Terminotix is a software development company that helps linguistic services and language service providers increase performance by automating and enhancing the translation process. Terminotix helps translators, revisers, coordinators and terminologists increase their productivity by up to 50% with a complete suite of products designed and optimized for language professionals.

Terminology Extraction as a Tool for MT Output Assessment and Improvement

The present paper proposes the use of a Terminology Recall Index (TRI) calculated on retaining nominal groups’ frequencies and stemming info only.

Though this paper proposes to demonstrate the utility of a TRI calculation between a human translated document and neural machine translated document, it also attempts to demonstrate that a broader use of the TRI calculation has many other surprising applications inside a linguistic service’s translation workflow.

Lecture Theatre
Thursday 21 at 15.15
Ayla Rigouts Terryn, Lieve Macken, Els Lefever, Robert Vander Stichele, Koen Vanneste and Joost Buysschaert (Univ/. of Ghent)

Ayla Rigouts Terryn is a PhD scholar at the LT3 language and translation technology team at Ghent University. She has a master in Translation from Antwerp University and researched translation revision competence at her alma mater. In 2015, she started at Ghent University to pursue her interests in natural language processing on the SCATE (Smart Computer-Aided Translation Environment) project. She currently holds a PhD scholarship from the Research Foundation – Flanders to study monolingual and multilingual automatic terminology extraction from comparable corpora. Her other research interests include medical translation and the difference between laypeople and specialists for terminology and translation tasks.

Lieve Macken is Assistant Professor at the Department of Translation, Interpreting and Communication at Ghent University (Belgium). She has strong expertise in multilingual natural language processing. Her main research focus is translation technology and more specifically the comparison of different methods of translation (human vs. post-editing, human vs. computer-aided translation), translation quality assessment, and quality estimation for machine translation. She is the operational head of the language technology section of the department, where she also teaches Translation Technology, Machine Translation and Localisation.

Els Lefever is an assistant professor at the LT3 language and translation technology team at Ghent University. She has a master in Linguistics and Literature (Romance languages) and holds a PhD in computer science from Ghent University on ParaSense: Parallel Corpora for Word Sense Disambiguation (2012). She started her career as a computational linguist at the R&D-department of Lernout & Hauspie Speech products. She has a strong expertise in machine learning of natural language and multilingual natural language processing, with a special interest for computational semantics, cross-lingual word sense disambiguation and multilingual terminology extraction. Currently, she supervises PhD research on the automatic extraction of topics, stance and argumentation from social media text, extracting terminology from comparable text, resolving ambiguous terms in cross-disciplinary collaboration and the automatic linking of medical lay and professional terminology to enhance comprehension of medical texts by patients. She is an executive board member of SIGLEX, the Special Interest Group on the Lexicon of the Association for Computational Linguistics and co-director of the Ghent Centre for Digital Humanities. She teaches Terminology and Translation Technology, Language Technology and Digital Humanities courses.

Joost Buysschaert is emeritus professor of the Department of Translation, Interpreting and Communication of Ghent University (Belgium), where he used to teach translation technology and medical translation (among other courses). He has remained active within the Department’s Terminology Centre (www.cvt.ugent.be) and continues to publish on terminology, translation tools and translation training. Among the terminology projects that he is involved in, is the MeSH Termbase Project on English and Dutch medical terminology (http://www.cvt.ugent.be/mesh.htm). He advises the company ivs iscientia on the use of translation tools.

Pilot Study on Medical Translations in Lay Language: Post-Editing by Language Specialists, Domain Specialists or Both?

Despite the rich history of research into medical translation, there is a notable lack of empirical studies on the best workflow for this task, especially in a modern translation setting involving post-editing of machine translation. This pilot study was conducted in preparation for a large translation project of medical guidelines for laypeople from Dutch into French. It is meant to shed light on how medical post-editing is best handled. How do medical specialists (doctors) versus language specialists (translators) perform on this task? How can their respective strengths lead to the highest quality translation? To gain more insight into these questions, errors in the machine translation output of medical guidelines were annotated and labelled.

Based on these annotations, the product of doctors’ and translators’ post-editing could be analysed and classified into necessary changes (mistakes that were correctly solved), underrevisions (mistakes that were not corrected during post-editing), overrevisions (new errors introduced during post-editing) and hyperrevisions (preferential changes made by the post-editor).

The results of this small-scale research illustrate the complexity of the task and reveal some surprising findings (e.g., doctors sometimes struggle with domain-specific terminology, and translators appear to be less efficient because they introduce many hyperrevisions).

Friday 22 at 9.15
Lecture Theatre
Jean Senellart (Systran)

Jean Senellart is the chairman of SYSTRAN, leading company in automated translation solutions for the last 50 years. Graduated from France Ecole Polytechnique and with a PhD on both language and computer science, Jean has been dedicating his work career to SYSTRAN, leading the development through the different generations of machine translation up to Neural Machine Translation in 2016. Cofounder of OpenNMT, the most popular open source frame dedicated to Neural Machine Translation, Jean is now striving to design solutions for breaking the language barrier and to provide language experts, language learners and language lovers with the amazing abilities coming from the neural technology.

Keynote speech: The Neural Revolution in the Translation Industry – 3-Year Retrospective and Future Directions Thursday 21 at 9.45
Lecture Theatre
Margita Šoštarić, Nataša Pavlović and Filip Boltužić (Univ. of Zagreb)

Margita Šoštarić is a recent graduate from the Faculty of Humanities and Social Sciences, University of Zagreb and currently works at Omega Software, a software development company in Zagreb. Her research interests span from the more theoretical approaches to language, such as cognitive linguistics, to the direct applications of language processing, such as machine translation.

Nataša Pavlović is an associate professor in the Faculty of Humanities and Social Sciences, University of Zagreb, where she teaches translation theory and practice. She holds a PhD in Translation and Intercultural Studies from the University Rovira i Virgili in Tarragona, Spain. Her research interests include translation process research, translator education, and translation technology, in particular machine translation and post-editing.

Filip Boltužić is a PhD researcher  in the Text Analysis and Knowledge Engineering Lab (TakeLab) in the Faculty of Electrical Engineering and Computing, University of Zagreb. His main research interests are natural language processing and argument mining. Prior to joining TakeLab, he worked at Zagrebačka banka Unicreditgroup as a data analyst and Amazon Data Services Ireland as a software development engineer.

Domain Adaptation for Machine Translation Involving a Low-Resource Language: Google AutoML vs. Fom-Scratch NMT System

Despite the advances in machine translation achieved with neural models, adaptation of such systems for specialist domains remains a challenge. The problem is particularly acute when it comes to low-resource languages.

Additionally, the computational resources and expertise needed to train neural models present barriers to entry for smaller translation companies and freelancers. In such cases, paid but affordable customization services such as Google Cloud AutoML might present a viable solution. In this study, domain adaptation using Google Cloud AutoML Translation is compared to a more traditional scenario, where several neural machine translation systems are trained from scratch using OpenNMT, an open-source toolkit for machine translation.

The from-scratch systems are trained using a larger out-of-domain English-Croatian dataset and a smaller in-domain English-Croatian dataset comprised of medical texts. The same in-domain data are used to customize Google Cloud AutoML Translation. The performance of the systems is compared using automatic and human evaluation methods. The resources, skills and time necessary to set up the examined systems are also discussed.

Thursday 21 at 12.30
Lecture Theatre
Maria Stasimioti and Vilelmini Sosoni (Ionian University)

Maria Stasimioti is a PhD candidate in the Department of Foreign Languages, Translation and Interpreting at the Ionian University. She holds a BA in Translation Studies and an MA in Theory and Didactics of Translation from the same university. She has been working as a freelance translator and proofreader since 2010. She has taught Computer-Assisted-Translation and English for Specific Purposes (ESP) at the Ionian University. She has also participated in the EU-funded project TraMOOC (Translation of Massive Open Online Courses, https://tramooc.eu/ ). Her research interests lie in the areas of Machine Translation (NMT, SMT), Computer-Assisted-Translation (CAT) and Post Editing (PE).

 

Dr Vilelmini Sosoni is Assistant Professor in the Department of Foreign Languages, Translation and Interpreting at the Ionian University in Corfu, Greece, where she teaches Legal and Economic Translation, EU texts Translation and Terminology, Translation Technology, Translation Project Management and Audiovisual Translation (AVT). In the past, she taught Specialised Translation in the UK at the University of Surrey, the University of Westminster and Roehampton University, and in Greece at the National and Kapodistrian University of Athens and the Institut Français d’ Athènes.  She also has extensive industrial experience having worked as translator, editor and subtitler.

She holds a BA in English Language and Linguistics from the National and Kapodistrian University of Athens, an MA in Translation and a PhD in Translation and Text Linguistics from the University of Surrey. Her research interests lie in the areas of Corpus Linguistics, Machine Translation (MT), Cognitive Studies, Translation of Institutional Texts and AVT. She is a founding member of the Research Lab “Language and Politics” of the Ionian University and a member of the “Centre for Research in Translation and Transcultural Studies” of Roehampton University. She has participated in several EU-funded projects, notably TraMOOC, Eurolect Observatory and Training Action for Legal Practitioners: Linguistic Skills and Translation in EU Competition Law, while she has edited several volumes and books on translation and published numerous articles in international journals and collective volumes.

Undergraduate Translation Students’ Performance and Attitude towards Machine Translation and Post-editing: Does Training Play a Role?

In an effort to meet the demands in speed and productivity, while keeping the cost low, the translation industry has turned to Machine Translation (MT) and Post-Editing (PE).

Nowadays, it is common practice to include MT in the translation workflow by using MT output as raw translation to be further post-edited by a translator (Lommel and DePalma, 2016). Yet, translators still approach PE with caution and skepticism and question its real benefits (Koponen 2012; Gaspari et al 2014; Moorkens 2018). In addition, attitudes to MT and PE seem to affect PE effort and performance (Witczak, 2016; Çetiner and İşisağ, 2019). Under that light, this study aims to investigate the attitudes and perceptions of undergraduate translation students towards MT and PE as well as their performance before and after they receive training in MT and PE.

Questionnaires are used to capture their attitudes and perceptions, while a human evaluation of their post-edited MT output is used to assess their performance and the quality of the post-edited texts. The analysis reveals a change in the students’ attitudes and perceptions; they report a more positive attitude toward MT and PE, they are more confident and faster, while they avoid over-editing.

Thursday 21 at 16.15
Lecture Theatre
Ales Tamchyna (Memsource)

Software Engineer, Head of the AI Team at Memsource. Ales joined Memsource in spring 2017 as the first member of the artificial intelligence (AI) team. He works on integrating smart solutions based on machine learning and AI in the Memsource platform. Before joining the company, he was an academic researcher. His primary research topic was machine translation.

Applying AI to NT and MT

In the translation industry, the disruptive effect of AI is not yet apparent. Machine learning/AI has traditionally been associated with one feature: machine translation (MT). It is true that with the recent advancements in neural MT, the output quality is inching closer to human translation. However, neural MT still makes serious mistakes and its quality can be upset by more complex sentences. More importantly, professional translation has different standards than simply passing for human translation; translations might require a specific style, consistent terminology, coherence across sentences and paragraphs, etc. But a translator’s goal is to convey the original meaning as closely as possible. They have to carefully navigate ambiguity and craft wording that best reflects the emphasis in the original text, ensuring that there can be no confusion about the meaning. Consider the severity of mistakes within legal texts or medical records; in marketing, a good translation can be the difference between a successful campaign and an international embarrassment. It’s clear that MT is not going to replace human translation anytime soon, if ever.

Thursday 21 at 12.00
Lecture Theatre
Andrzej Zydroń (XTM Int.)

Andrzej Zydroń (MBCS CITP) is one of the leading IT experts on Localization and related Open Standards. Zydroń sits/has sat on the following Open Standard Technical Committees:

  1. LISA OSCAR GMX
  2. LISA OSCAR xml:tm
  3. LISA OSCAR TBX
  4. W3C ITS
  5. OASIS XLIFF
  6. OASIS Translation Web Services
  7. OASIS DITA Translation
  8. OASIS OAXAL
  9. ETSI LIS
  10. DITA Localization
  11. Interoperability Now!
  12. Linport

Zydroń has been responsible for the architecture of the essential word and character count GMX-V (Global Information Management Metrics eXchange) standard, as well as the revolutionary xml:tm standard which will change the way in which we view and use translation memory. Zydroń is also head of the OASIS OAXAL (Open Architecture for XML Authoring and Localization technical committee.

Zydroń has worked in IT since 1976 and has been responsible for major successful projects at Xerox, SDL, Oxford University Press, Ford of Europe, DocZone and Lingo24.

De-demonizing AI

AI has garnered much hype over the past few years. Andrzej Zydroń provides a realistic definition of AI: what is intelligence; how can it be defined; what is the mathematical basis for intelligence, as well as detailing the theoretical limitations of AI and what is actually achievable. The presentation will detail the actual practical potential of AI as well as its limitations and pitfalls when human beings interact with AI systems.

Thursday 21 at 11.30
Lecture Theatre