
Dr Elena Davitti
Academic and research departments
Centre for Translation Studies, School of Literature and Languages, Faculty of Arts and Social Sciences.About
Biography
I am an Associate Professor in Translation Studies with expertise in interpreting, both conference and dialogue. I am also Programme Leader of the MA Interpreting (Multilingual pathway) and MA Translation and Interpreting offered by the Centre for Translation Studies (CTS) where I am based. I hold a PhD in Translation and Intercultural Studies from the University of Manchester and an MA in Conference Interpreting from the University of Bologna at Forlì. Before joining Surrey in 2013, I practised as a freelance interpreter and translator and worked as interpreter trainer at different universities both in the UK and in Italy, such as the University of Leeds, University of Birmingham, University of Macerata and UNINT, Rome. I am currently working on hybrid modalities at the crossroads of traditional disciplines such as translation, interpreting , subtitling, with a particular interest in real-time speech-to-text communication across languages.
Areas of specialism
University roles and responsibilities
- Programme Leader of MA Interpreting (Multilingual Pathway)
- Programme Leader of MA Translation and Interpreting
My qualifications
Affiliations and memberships
News
In the media
ResearchResearch interests
My research interests revolve around technology-enabled methods, modalities and practices of multilingual spoken communication, namely:
- Real-time speech-to-text transfer across languages via:
- interlingual respeaking, as a hybrid modality combining advances in speech recognition technology with human interpreting and subtitling skills to improve access to multilingual audiovisual content for a wider audience, its impact on language professional and traditional practices; required skills, abilities and traits; output quality; broader socio-economic impact;
- alternative (semi-)automated workflows integrating speech recognition and/or machine translation to deliver the same service, especially comparison of the fit-for-purposeness in different contexts; human role in increasingly technologised workflow;
- Video-mediated interpreting, with particular emphasis on communicative dynamics, impact on interpreters and participation dynamics, interpreting quality, traditional professional practice, professionalisation and interpreter education;
- Multimodal and interactional approaches to interpreter-mediated interaction, both face-to-face (particularly in educational and medical settings) and technology-mediated.
Research projects
The recent worldwide audiovisual content boom has led to an ever-increasing demand for such content to be made accessible in real-time, in different languages, and for a wide audience in a variety of settings, including television, conferences and other live events (e.g. lectures, museum tours, business meetings, medical appointments). The SMART project (Shaping Multilingual Access through Respeaking Technology), funded by the ESRC UK (Economic and Social Research Council UK, ES/T002530/1, 2020-2023), addresses the urgent challenge to deliver high-quality real-time speech-to-text services across languages by exploring a new practice, interlingual respeaking. SMART will investigate how this technique can be fine-tuned to best produce subtitles in a different language, and the impact this may have on society.

Research interests
My research interests revolve around technology-enabled methods, modalities and practices of multilingual spoken communication, namely:
- Real-time speech-to-text transfer across languages via:
- interlingual respeaking, as a hybrid modality combining advances in speech recognition technology with human interpreting and subtitling skills to improve access to multilingual audiovisual content for a wider audience, its impact on language professional and traditional practices; required skills, abilities and traits; output quality; broader socio-economic impact;
- alternative (semi-)automated workflows integrating speech recognition and/or machine translation to deliver the same service, especially comparison of the fit-for-purposeness in different contexts; human role in increasingly technologised workflow;
- Video-mediated interpreting, with particular emphasis on communicative dynamics, impact on interpreters and participation dynamics, interpreting quality, traditional professional practice, professionalisation and interpreter education;
- Multimodal and interactional approaches to interpreter-mediated interaction, both face-to-face (particularly in educational and medical settings) and technology-mediated.
Research projects
The recent worldwide audiovisual content boom has led to an ever-increasing demand for such content to be made accessible in real-time, in different languages, and for a wide audience in a variety of settings, including television, conferences and other live events (e.g. lectures, museum tours, business meetings, medical appointments). The SMART project (Shaping Multilingual Access through Respeaking Technology), funded by the ESRC UK (Economic and Social Research Council UK, ES/T002530/1, 2020-2023), addresses the urgent challenge to deliver high-quality real-time speech-to-text services across languages by exploring a new practice, interlingual respeaking. SMART will investigate how this technique can be fine-tuned to best produce subtitles in a different language, and the impact this may have on society.

Supervision
Postgraduate research supervision
I welcome enquiries from prospective PhD candidates with projects in the following areas:
- Spoken language interpreting (all modes and modalities)
- Hybrid modalities of speech-to-text transfer via speech recognition, especially interlingual respeaking
- Semi-/fully-automated workflows for interlingual speech-to-text in real time
- Interpreting technologies
- Video-mediated interpreting (all modes and configurations)
- Multimodal approahches to interpreter-mediated interaction
- Micro-analytical and empirical analysis of communicative and interactional dynamics in interpreting
- Interpreter education, upskilling and professionalisation
Current PhD students
Main supervisor
- Radić, Željko .Integrating speech recognition technology and human subtitling skills for the translation of interlingual subtitles
- Madell, Soumely. Technology-enhanced multilingual healthcare communication in the NHS maternity setting
Co-supervisor
- Rodríguez González, Eloy. The use of speech recognition in remote simultaneous interpreting
- Saeed, Muhammad Ahmed. The role of presence in remote simultaneous interpreting
- Zhang, Wei. Patient-centred approaches in medical interpreting.
Completed PhD project
- Gabrych, Marta. (2019). Quality Assessment of Interpreting in Polish-English Police-Suspect Interviews.
- Carpi, Beatrice (2018). Systematizing the Analysis of Songs in Stage Musicals for Translation: A Multimodal Model Based on Themes.
- Al-Jabri, Hanan (2017). TV Simultaneous Interpreting of Arabic Presidential Speeches into English During the Arab Spring.
Publications
This paper examines the work of project managers in two UK-based translation companies. Drawing on participant observation, interviews, and artifacts from field sites, our analysis focuses on the ways in which trust is developed and maintained in the relationships that project managers build, on the one hand, with the clients who commission them to undertake translation projects, and, on the other, with freelance translators who perform the translation work. The project manager’s ability both to confer and to instill trust is highlighted as key to the successful operation of the company. Conceptualizing trust as a dynamic process, we consider what this process of trusting entails in this context: positive expectations visà-vis the other parties; willingness to expose oneself to vulnerabilities; construction of bases for suspending doubts and uncertainties (leaps of faith). We observe the important role of communication and discursive strategies in building and maintaining trust and draw conclusions for translator education.
In the last two decades, Dialogue Interpreting (DI) has been studied extensively through the lenses of discourse analysis and conversation analysis. As a result, DI has been recognised as an interactional communicative event, in which all the participants jointly and actively collaborate. Nevertheless, most of these studies focused merely on the verbal level of interaction, whereas its multimodal dimension has not received much attention so far, and the literature on this subject is still scarce and dispersed. By analysing and comparing two sequences, taken from a corpus of face-to-face interpreter-mediated encounters in pedagogical settings, this study aims at showing how multimodal analysis can contribute to a deeper understanding of the interactional dynamics of DI. In particular, the paper shows how participants employ multimodal resources (gaze, gesture, body position, proxemics, object manipulation) to co-construct different participation frameworks throughout the encounters, and how the “ecology of action” (i.e., the relationships between the participants and the surrounding environment) influences the development of interaction.
This paper presents the key findings of the pilot phase of SMART (Shaping Multilingual Access though Respeaking Technology), a multidisciplinary international project focusing on interlingual respeaking (IRSP) for real-time speech-to-text. SMART addresses key questions around IRSP feasibility, quality and competences. The pilot project is based on experiments involving 25 postgraduate students who performed two IRSP tasks (English-Italian) after a crash course. The analysis triangulates subtitle accuracy rates with participants’ subjective ratings and retrospective self-analysis. The best performers were those with a composite skillset, including interpreting/subtitling and interpreting/subtitling/respeaking. Participants indicated multitasking, time-lag, and monitoring of the speech recognition software output as the main difficulties; together with the great variability in performance, personal traits emerged as likely to affect performance. This pilot lays the conceptual and methodological foundations for a larger project involving professionals, to address a set of urgent questions for the industry.
In recent years, conversation analysts have developed a growing interest in the Applied branch of Conversation Analysis (CA). Authors such as Antaki, Heritage and Richards and Seedhouse have explored the practical applications of CA in institutional contexts, to shed light on their dynamics and to suggest improvements in the services provided. On the other hand, over the past two decades, interactionally oriented methodologies have been successfully applied to the study of interpreter-mediated interaction. Nevertheless, the potential of CA for interpreter training has not been fully explored, especially with regard to the impact of multimodal semiotic resources (gaze, gesture, posture) on triadic communication. This paper illustrates the results of an exploratory study in Applied CA conducted within a postgraduate interpreting module at an Italian university. Four different extracts of interpreter-mediated encounters, video-recorded in real-life settings, were submitted to the students in order to test their reactions, guide them in analysis and raise their awareness of the problems and challenges posed by real-life scenarios. Through the triangulation of findings from recorded classroom discussion and questionnaires, the present paper discusses the implications of using CA in interpreter education.
Video Remote Interpreting (VRI) is a modality of interpreting where the interpreter interacts with the other parties-at-talk through an audiovisual link without sharing the same physical interactional space. In dialogue settings, existing research on VRI has mostly drawn on the analysis of verbal behaviour to explore the complex dynamics of these ‘triadic’ exchanges. However, understanding the complexity of VRI requires a more holistic analysis of its dynamics in different contexts as a situated, embodied activity where resources other than talk (such as gaze, gestures, head and body movement) play a central role in the co-construction of the communicative event. This paper draws on extracts from a corpus of VRI encounters in collaborative contexts (e.g. nurse-patient interaction, customer services) to investigate how specific interactional phenomena which have been explored in traditional settings of dialogue interpreting (e.g. turn management, dyadic sequences, spatial management) unfold in VRI. In addition, the paper will identify the coping strategies implemented by interpreters to deal with various challenges. This fine-grained, microanalytical look at the data will complement the findings provided by research on VRI in legal/adversarial contexts and provide solid grounds to evaluate the impact of different moves. Its systematic integration into training will lead to a more holistic approach to VRI education.
Research in Dialogue Interpreting (DI) has traditionally drawn on qualitative analysis of verbal behaviour to explore the complex dynamics of these ‘triadic’ exchanges. Less attention has been paid to interpreter-mediated interaction as a situated, embodied activity where resources other than talk (such as gaze, gestures, head and body movement, proxemics) play a central role in the co-construction of the communicative event. This article argues that understanding the complexity of DI requires careful investigation of the interplay between multiple interactional resources, i.e. verbal in conjunction with visual, aural, embodied and spatial meaning-making resources. This call for methodological innovation is strengthened by the emergence of video-mediated interpreting, where interacting via screens without sharing the same physical space adds a further layer of complexity to interactional dynamics. Drawing on authentic extracts from interpreter-mediated interaction, both face-to-face and video-mediated, this article problematizes how the integration of a multimodal perspective into qualitative investigation of interpreter-mediated interaction can contribute to the advancement of our understanding of key interactional dynamics in DI and, in turn, broaden the scope of multimodality to include new, uncharted territory.
Research on dialogue interpreting shows that interpreters do not simply convey speech content, but also perform crucial coordinating and mediating functions. This descriptive study, which is based on PhD research conducted at the University of Manchester, explores the activity of qualified dialogue interpreters in three video-recorded parent-teacher meetings involving immigrant mothers. English and Italian are the languages used, the meetings having taken place in the UK (one case) and Italy (two cases). The study focuses on interpreters' handling of evaluative assessment, in many cases introduced by them in the target speech as an "upgrading rendition". Transcribed extracts are examined in a micro-analytical perspective, the dynamics of each actor's (dis)engagement towards interlocutors being studied in relation to gaze patterns annotated by dedicated software. Results show that the interpreter actively promotes alignment between the parties; however, s/he often does so by emphasising positive considerations to the mother. The outcome of this approach is that the mother accepts, but is not encouraged to co-construct a negotiated solution: she is assimilated, not empowered. © John Benjamins Publishing Company.
We report on a study evaluating the educational opportunities that highly multimodal and interactive Virtual Learning Environments (VLE) provide for collaborative learning in the context of interpreter education. The study was prompted by previous research into the use of VLEs in interpreter education, which showed positive results but which focused on preparatory or ancillary activities and/or individual interpreting practice. The study reported here, which was part of a larger project on evaluating the use of VLEs in educating interpreters and their potential clients, explored the affordances of a videoconferencing platform and a 3D virtual world for collaborative learning in the context of dialogue interpreting. The participants were 13 student-interpreters, who conducted role-play simulations in both environments. Through a mix of methods such as non-participant observation, reflective group discussions, linguistic analysis of the recorded simulations, and a user experience survey several dimensions of using the VLEs were explored including the linguistic/discursive dimension (interpreting), the interactional dimension (communication management between the participants), the ergonomic dimension (human-computer interaction) and the psychological dimension (user experience, sense of presence). Both VLEs were found to be capable of supporting situated and autonomous learning in the interpreting context, although differences arose regarding the reported user experience.
Remote simultaneous interpreting (RSI) draws on Information and Communication Technologies to facilitate multilingual communication by connecting conference interpreters to in-presence, virtual or hybrid events. Early solutions for RSI involved interpreters working in interpreting booths with ISOstandardised equipment. However, in recent years, cloud-based solutions for RSI have emerged, with innovative Simultaneous Interpreting Delivery Platforms (SIDPs) at their core, enabling RSI delivery from anywhere. SIDPs recreate the interpreter's console and work environment (Braun 2019) as a bespoke software/videoconferencing platform with interpretation-focused features. Although initial evaluations of SIDPs were conducted before the Covid-19 pandemic (e.g., DG SCIC 2019), research on RSI (booth-based and software-based) remains limited. Pre-pandemic research shows that RSI is demanding in terms of information processing and mental modelling (Braun 2007; Moser-Mercer 2005), and suggests that the limited visual input available in RSI constitutes a particular problem (Mouzourakis 2006; Seeber et al. 2019). Besides, initial explorations of the cloud-based solutions suggest that there is room for improving the interfaces of widely used SIDPs (Bujan and Collard 2021; DG SCIC 2019). The experimental project presented in this paper investigates two aspects of SIDPs: the design of the interpreter interface and the integration of supporting technologies. Drawing on concepts and methods from user experience research and human-computer interaction, we explore what visual information is best suited to support the interpreting process and the interpreter-machine interaction, how this information is best presented in the interface, and how automatic speech recognition can be integrated into an RSI platform to aid/augment the interpreter's source-text comprehension.
This book examines how researchers of discourse analysis could best disseminate their work in real world settings. The chapters include studies on spoken and written discourse using various analysis techniques, and the authors discuss how they could best engage professional practice in their work. Techniques used include Conversation Analysis in combination with other methods, genre analysis in combination with other methods, and Critical Discourse Analysis. Contributions are loosely grouped by setting and include the following settings: workplace and business; education; private and public; and government and media. The volume aims to link the end of research and the onset of praxis by creating collaboration with the places of practice, helping analysts to move forward with ideas for dissemination, collaboration and even intervention. The book will be of interest to all researchers conducting discourse analysis in professional settings.
This volume focuses on multimodality in various communicative settings, with special attention to how non-verbal elements reinforce and add meaning to verbal expressions. The first part of the book explores issues related to the use of multimodal resources in educational interactions and English language classroom teaching, also involving learners with disabilities. The second part, on the other hand, investigates multimodality as a key component of communication that takes place in different specialized domains and genres. The book reflects a variety of methodological approaches that are grounded in both quantitative and qualitative techniques. These include multimodal discourse analysis, multimodal transcription, and multimodal annotation software capable of representing the interplay of different semiotic modes, such as speech, intonation, direction of gaze, facial expressions, gestures and spatial positioning of interlocutors.
The Routledge Encyclopedia of Interpreting Studies is the authoritative reference for anyone with an academic or professional interest in interpreting. Drawing on the expertise of an international team of specialist contributors, this single-volume reference presents the state of the art in interpreting studies in a much more fine-grained matrix of entries than has ever been seen before. For the first time all key issues and concepts in interpreting studies are brought together and covered systematically and in a structured and accessible format. With all entries alphabetically arranged, extensively cross-referenced and including suggestions for further reading, this text combines clarity with scholarly accuracy and depth, defining and discussing key terms in context to ensure maximum understanding and ease of use. Practical and unique, this Encyclopedia of Interpreting Studies presents a genuinely comprehensive overview of the fast growing and increasingly diverse field of interpreting studies.
In the last two decades, empirical research has shed light on the interactional dynamics of Dialogue Interpreting (DI). Nevertheless, it remains unclear how the results of such research can be effectively integrated in interpreter education. This paper outlines a semester long module, in which research on DI is employed for teaching purposes. During the module, students are introduced to relevant literature and exposed to different case studies of interpreter-mediated interaction, based on authentic data. The aim is to create an understanding of the interpreter ’s role and conduct in a variety of communicative situations, and help students identify the challenges that may arise in interpreter-mediated interaction. Implications for current codes of conduct are also discussed.
This chapter reports the key findings of the European AVIDICUS 3 project,1 which focused on the use of video-mediated interpreting in legal settings across Europe. Whilst judicial and law enforcement authorities have turned to videoconferencing to minimise delays in legal proceedings, reduce costs and improve access to justice, research into the use of video links in legal proceedings has called for caution. Sossin and Yetnikoff (2007), for example, contend that the availability of financial resources for legal proceedings cannot be disentangled from the fairness of judicial decision-making. The Harvard Law School (2009: 1193) warns that, whilst the use of video links may eliminate delays, it may also reduce an individual’s “opportunity to be heard in a meaningful manner”. In proceedings that involve an interpreter, procedural fairness and “the opportunity to be heard in a meaningful manner” are closely linked to the quality of the interpretation. The use of video links in interpreter-mediated proceedings therefore requires a videoconferencing solution that provides optimal support for interpreting as a crucial prerequisite for achieving the ultimate goal, i.e. fairness of justice. Against this backdrop, the main aim of AVIDICUS 3 was to identify institutional processes and practices of implementing and using video links in legal proceedings and to assess them in terms of how they accommodate and support bilingual communication mediated through an interpreter. The focus was on spoken-language interpreting. The project examined 12 European jurisdictions (Belgium, Croatia, England and Wales, Finland, France, Hungary, Italy, the Netherlands, Poland, Scotland, Spain and Sweden). An ethnographic approach was adopted to identify relevant practices, including site visits, in-depth and mostly in-situ interviews with over 100 representatives from different stakeholder groups, observations of real-life proceedings, and the analysis of a number of policy documents produced in the justice sector. The chapter summarises and systematises the findings from the jurisdictions included in this study. The assessment focuses on the use of videoconferencing in both national and cross-border proceedings, and covers different applications of videoconferencing in the legal system, including its use for links between courts and remote participants (e.g. witnesses, defendants in prison) and its use to access interpreters who work offsite (see Braun 2015; Skinner, Napier & Braun in this volume).
Additional publications
Korybski, T. E. Davitti, C. Orasan, and S. Braun (2022) A Semi-Automated Live Interlingual Communication Workflow Featuring Intralingual Respeaking: Evaluation and Benchmarking, Proceedings of the Language Resources and Evaluation Conference (LREC), June 2022, European Language Resources Association, pp. 4405--4413