Most academic research on Audiovisual Translation and specifically on Media Accessibility has focused on content creation. If media content today can be compared to the visible tip of the iceberg, technology represents the unknown hidden part. Accessibility services are epigonic to media content, presenting an intimate relationship between technology for delivery and technology for consumption. Technology allows not only for the production of media content and access services, but also elicits new formats and interactions.
This paper focuses on accessibility services delivery. The first part provides a brief on state-of-the-art accessibility apps available in the market today, including emerging trends and market uptake. Section 2 presents audience needs away from the classical medical classification of assistive technologies and an analysis of existing accessibility apps. Section 3 presents a novel solution that is currently available in the market for transmedia storytelling. The new app has been designed to provide personalised access, away from one-size-fits-all services.
Accessibility, distribution, personalisation, large audience, live events.
Smartphones have become one of the most popular ways to consume media content (Casetti and Sampietro 2013, Weinel and Cunningham 2015) , with particular focus on movies (Hessels 2017). The distribution of media content in a live condition and the associated accessibility files presents an interesting challenge, with far too many variables for a single, comprehensive solution (Oncins et al. 2013). A Live condition becomes even more complex when audience needs regarding interaction are taken into consideration (Linke-Ellis 2012). No single solution will work for all audiences. Nevertheless, a personalised modular system that is adaptable to user requirements will facilitate the comprehensive delivery of live accessibility services.
In the past 10 years many apps have been designed to cater for accessibility services and to subjugate communication barriers for persons with disabilities (Oncins 2014). This follows the EU legislation towards accessibility with the updated Audiovisual Media Services Directive 2018/1808 and the European Accessibility Act, Directive (EU) 2019/882, and complies with United Nations Convention on the Rights of Persons with Disabilities (UN CRPD 2008). However, accessibility in live events remains problematic.
The main issues are threefold: the cost of the event, the low return on investment and resilience. It is also an issue from a technological standpoint in terms of service distribution or display, neither of which are usually available in open air or closed conditions.
Nowadays, accessibility services at live events are provided in three ways:
It should be highlighted that these three provisions often interact in order to provide an additional accessibility service. An example would be an opera house that regularly offers subtitling services but requires sign language interpretation or audio description for one specific performance. In this case a service provider company or freelance expert will be contracted. The delivery a new ad-hoc service for only one live performance may imply ephemeral and costly technological solutions.
According to the latest data provided from the European Health and Social Integration Survey (EHSIS 2012a), there are more than 70 million people with a disability in the European Union (EU), which means a disability rate of 17.6% in the total population of the EU27 aged 15 and over. In addition, the process of ageing is increasing the number of EU citizens deemed as having some type of impairment.
According to the data, there are an estimated 84 to 107 million people with a disability in the EU. The reason for this range can be attributed to prevailing differences in culture in the Member States along with differing definitions of disability, leading to inaccurate measurements.
The European health and social integration survey (EHSIS 2012b) presents an overview of the special assistance needed by people with disabilities aged 15 and over in the EU, in order to subjugate limitations or barriers. The data shows the areas in which these people face barriers regarding social participation. As shown in Figure 1, leisure pursuits is the area where the largest proportion of people with disabilities reported limitations.
Figure 1. Share of disabled persons aged 15 and over reporting a disability in the specified life areas, by sex, EU-27, 2012 (as a % of persons reporting a disability in at least one area)
The report includes a classification according to “a health component to identify long-standing health problems and activity difficulties” (EHSIS 2012b: 1). Based on a medical model, this audience classification is the standard approach in experimental Audiovisual Translation research (Agulló et al. 2018, Orero and Tor-Carroggio 2018, Tor-Carroggio and Orero 2019). Research in this field is performed in two independent silos dealing with two accessibility services — subtitling and audio description for people with hearing loss and sight loss respectively (Agulló et al. 2018). In the former, research has departed from a variable number of deaf and hard of hearing users. Studies in same language subtitles started with Neves (2007) and Arnáiz-Uzquiza (2012) who defined the target audience as people with hearing loss. The same population is studied in experiments by Romero-Fresco (2009, 2010, 2013, 2015), Bartoll (2004, 2008, 2012), Bartoll and Martínez-Tejerina (2010), Perego et al. (2010), Pereira (2010), Szarkowska (2011), Szarkowska et al. (2011 and 2016), Miquel-Iriarte (2017) and Tsaousi (2017). All these studies have the same objective: understanding same language subtitle processing that has been created for people with hearing loss. The variables go from the reproduction speed, the subtitle position and the format. None of the previous studies tested the participants’ reading ability, nor how the semantic value of words may determine comprehension. The other popular media accessibility service is audio description. In this case the service targets persons with sight loss. Audio description (AD) deals with sound track perception and seeks to convey and compliment the message of the audiovisual text. This complex audio narrative is at the centre of studies by Udo and Fels (2009), Szarkowska (2011), Chmiel and Mazur (2012a), Szarkowska and Jankowska (2012), Fryer and Freeman (2012, 2014), Romero-Fresco and Fryer (2013), Fresno et al. (2014), Walczak and Rubaj (2014), Szarkowska and Wasylczyk (2014), Fernández-Torné and Matamala (2015), Walczak and Fryer (2017, 2018), and three experimental PhD dissertations: Cabeza-Cáceres (2013), Fryer (2013), and Walczak 2017 (framed within the EU-funded project HBB4ALL1). Other research results from large projects such as DTV4ALL2, ADLAB3, the Pear Tree Project (Orero 2008, Chmiel and Mazur 2012b), AD-Verba (Chmiel and Mazur 2011) and OpenArt (Szarkowska et al. 2016). In all of these studies the objective is to understand the reception of audio description as different stimuli are presented. The variables go from the narration speed, the intonation, the explicitation, the enjoyment and the use of synthetic voices, among other. None of the previous studies tested the participants’ hearing ability. The target audience is persons with sight loss; however, the final audio description service is an audio track, which relates more to hearing issues than sight issues.
There is no dispute that accessibility services are developed for people with disabilities. However, Audiovisual Translation studies on Media Accessibility still use medical conditions towards profiling their audience demographics: hearing loss or sight loss. Neither subtitling nor audio description will restore the hearing or sight of its audience. In the study of reading subtitles, the focus has been on reading skills. Profiling should be based on reading for legibility, readability and comprehension. To date, comprehension of reading subtitles has been related to the semantic complexity of words or concepts and is an interesting avenue for further research and development (Perego et al. 2010). Research on reading speed or live subtitling delay are measured according to the number of characters per second or words per minute. These measurements take syntax and the average length of words into consideration as they are language dependent, as opposed to dependent on the reception of the content. This was the case in a study by González-Iglesias and Martínez Pleguezuelos (2010) on gender violence in institutional TV adverts which provide accessibility services in Spain. Profiling people with hearing loss as the target group for gender violence and offering verbatim subtitles with the associated reading speed may lead to the belief that only people with hearing loss are in need of help and advice related to gender violence. Verbatim subtitles are an institutional requirement of hearing loss associations.
This traditional, medical audience classification has led to the design of solutions addressed to either one disabled group or the other (Sánchez Sierra and Selva Roca de Togores 2012). This can be seen in the table below (Table 1) where 23 apps have been organised according to the access service they provide and by service distribution: live, pre-recorded or recorded conditions.
Table 1. Existing media accessibility apps according to services
If we take a closer look at the apps described in the table that deliver pre-recorded audiovisual (AV) content, seven offer multilingual subtitles and eight offer subtitles for the deaf and hard of hearing (SDH). It should be highlighted that not all apps offer both types of subtitles. While the former usually provides subtitles in at least two different languages, the latter offers subtitles that include non-verbal information and the use of colours, which aid character identification. Regarding ADs of pre-recorded AV content, 11 apps cover this service, three apps provide sign language and one app offers dubbing. Out of all the apps, six include multilingual subtitles, SDH or AD whereas only Moviereading integrates all three services. This again shows audience segmentation through clinical requirements versus communicative requirements. Designed technology has focused on accessibility services that cater to the needs of users with disabilities, instead of approaching a wider notion of end-user capabilities (Agulló et al. 2018, Orero and Tor-Carroggio 2018).
Regarding semi-live events, nine apps offer SDH and six offer multilingual subtitles. Regarding ADs, 10 apps have it covered, eight of which integrate two main services: SDH and audio-description. Only one app provides sign language. In terms of live events, we find a clear requirement for apps that will provide all three accessibility services. Only five apps integrate SDH and none provide live interlingual subtitles. In live events the most commonly used techniques are respeaking and typing (Oncins 2014). In the case of AD, the service is covered by four of the apps.
It is important to mention that three apps offer sign language, namely Moviereading, Actiview and Whatscine, of which only the last offers the service in semi-live and live conditions. This may be due to the fact that the streaming of videos is subject to latency problems, which continue to be a major challenge for the audiovisual industry. The introduction of 5G connections in the near future might be a major step towards problem mitigation if not a solution.
In terms of technologies used to synchronise and deliver accessibility services, there are two main trends; the first, the integration of artificial intelligence (AI) technology, which improves speech recognition, i.e. Whatscine is used in semi-live events to automatically synchronise the AV content with the accessibility services. It is important to highlight that in semi-live events, accessibility contents are initially uploaded to the platform and then delivered manually. The use of AI technology is employed to automate this delivery process. Still, some of the main challenges in terms of synchronisation and voice recognition are presented when actors skip lines from the scripts, make improvisations, increase the speed, alter the prosody, have strong accents, use slang or when there is background noise.
The second notable trend is an increase in the use of the audio channel in addition to the AD service, i.e. dubbing or voice over, offered by GalaPro for live and semi-live events, and TheaterEars for pre-recorded events. Speekie and GalaPro offer live AD services for conferences and other types of events for which there is no script available. Speekie also allows live commentary and real-time translation at conferences and speeches.
Interestingly, some of the apps that include multi-language subtitles offer services beyond the scope of assistive technologies, such as alternative content related to the AV content, in order to appeal to a broad target audience. However, apps covering SDH and ADs are limited to content which exclusively relates to the available performances that provide accessibility services.
3. From Assistive Technologies to Design for All
As observed in the previous section, to date, most apps intended for the use of accessibility services following the medical model of disability are categorised as assistive technologies (Orero and Tor-Carroggio 2018). The UN World Health Organisation published the International Classification of Functioning, Disability and Health (ICF) in 2001. The ICF was intended to complement its sister classification system, the International Classification of Diseases (ICD) (Peterson and Elliott 2008). The ICF Model sees disability as the complex context of an individual, taking institutional and societal factors into account (Dubois and Trani 2009). It is operationalised through the World Health Organization Disability Assessment Schedule II (WHODAS II) and offers a taxonomy of disabilities that serves as little more than a classification (Dubois and Trani 2009). Ellis (2016) pointed out the underlying difference between disability and impairment, and the need for an understanding of end user capabilities, beyond disability. Confronting the limitations of existing approaches leads to the mainstreaming of accessibility and the inclusion of user profiles beyond persons with disabilities. As pointed out by Agulló et al., “The lack of capabilities linked to disability may have less impact than the lack of capabilities linked to other aspects such as technology” (2018: 198). This concept has also been suggested as the way forward by the UN agency International Telecommunication Union (ITU):
Besides the more commonly used “medical model of disability”, which considers disability “a physical, mental, or psychological condition that limits a person’s activities”, there is a more recent “social model of disability”, which has emerged and is considered a more effective or empowering conceptual framework for promoting the full inclusion of persons with disabilities in society. Within the social model, a disability results when a person who (a) has difficulties reading and writing; (b) attempts to communicate, yet does not understand or speak the national or local language, and (c) has never before operated a phone or computer and attempts to use one – without success. In all cases, disability has occurred, because the person was not able to interact with his or her environment. (2017:2)
The ITU recommends leaving behind the clinical model of disability and to embrace a social model. Some research results in Audiovisual Translation have already indicated the need for change in audience classification. Romero-Fresco’s (2015) findings on reading subtitles linked readability to educational background rather than to their hearing impairment. Miquel-Iriarte (2017), Neves (2018) and Agulló et al. (2018) also raised the issue, beyond the audience hearing needs, moving “from accessibility to usability for diverse audiences” (Agulló et al. 2018: 198). In the same line, AD can be beneficial for a much broader audience than the visually impaired. It has proven to be effective for language learning, writing skills development and improvement of learning outcomes (Jankowska 2019, Walczak 2016).
The need for a comprehensive tool, designed for accessing cultural events in live conditions from a capability approach within Design for All principles was the context for MyAcc, the new inclusive multiservice app.
The departure point for the development of MyAcc had previous shortcomings regarding the live delivery of accessibility services, mainly relating to cost. A precursor for MyAcc was developed at Universitat Autònoma de Barcelona (UAB) in an attempt to deliver accessibility during live lectures, institutional academic acts, theatre performances and movies screened at the cinema. According to the UAB figures from 20194, the university has an average of 30,000 students per year including undergraduate, master and lifelong learning students. In turn UAB has an average of 4,000 foreign students. Teaching and institutional acts could be delivered in any of the three official UAB languages: Catalan, English and Spanish. This situation creates real demand for language accessibility, where profiling regarding numbers shifts from people with sensorial disabilities to people with language disabilities. UAB started an active accessibility campaign, organising accessible activities and purchasing state of the art equipment for the production, distribution and delivery of said activities (Oncins et al. 2013). The outcome of this accessibility campaign lead to the opening of a campus Accessibility Service that is also responsible for the production of accessibility service content. The price of the production of content in three languages posed concerns about its financial viability, leading to alternative production avenues, such as hiring external providers. As soon as one space was made accessible, other spaces requested accessibility, which created additional funding issues and the need for a solution for all on-campus UAB spaces. This lead to further scrutiny of the accessibility delivery system, due to it being independent of the venue equipment. Taking this into account, requirements were defined using a new, communicative capabilities classification of user profile, departing from the medical model. The new app was designed to fulfil four main principles:
Within this context, the app was designed to integrate all accessibility services in a single web platform with a vision beyond disability. Two conditions required by the app are defined as follows: the environment where the media content is consumed and the alternative communication of the visual, audio, and linguistic components of the media content. The integration of technologies such as Text-to-Speech, Speech-to-Text and Machine Translation has been considered part of the alternative communication, catering to the needs of an increasingly diverse audience. Finally, the full accessibility chain has been taken into account according to the Design for all principles.
MyAcc is a digital solution for transmedia storytelling and personalised communication. It works for one person and very large audiences. MyAcc is able to deliver in real-time, both for live and pre-recorded audio, video, texts, and images. Content delivery is personalised according to user profiling and communicative needs (see Figure 2).
Figure 2. Available services included in the app
The use of the personal smartphone and its capabilities implies no hardware dependency. The use of the personal smartphone removes the need for the end user to learn any functionality. User interaction is intuitive through an accessible menu whereby the services are chosen. The following accessibility services are included:
More than one service can be selected from the menu without interfering with the needs of the user. In the same auditorium, different users have different needs and/or more than one need. For instance; if a user selects English language, subtitles and audio description, the delivered services will be audio description and audio subtitles in English. A second user can select Spanish language and subtitles and in doing so the delivered service will be in Spanish. It addresses the coverage of a wide notion of accessibility services:
The new heterogeneity of the audience profile and audience attitudes demands new approaches in Media Accessibility. In addition, new consumer attitudes have to be taken into account. Data show that 80% of people that use subtitles are not deaf and use “common accessibility tools” to overcome problems relating to their environments. Furthermore, voice recognition services are expanding to new horizons, primarily aimed at the blind and people with sight-loss. It could be stated that voice and voice-enabled devices are increasing at a fast rate and voice-activated speakers have become part of people’s routines (Kleinberg 2018). The use of these new technologies is expected to increase in the coming years and the human-computer interaction (HCI) through voice recognition will become normalised. It is therefore expected from an end-user point of view that speaking and receiving audio content from an app will be accepted. This further extends the accessibility to a broader audience.
Attending a cultural event starts before the audience is seated. Buying the ticket and finding available accessible transport to and from the venue are both steps in the process towards a fully accessible experience. How to get to the assigned seat may also be a challenge. According to Universal Design or Design for All principles, the accessibility chain starts when people make the decision to attend a venue, followed by how to get to the venue, the experience itself, a safe return home and subsequent feedback to allow end-users to provide their insights into the experience. If any element in the chain is broken then accessibility cannot be considered as granted (Orero 2017).
MyAcc has been conceived to create new communication channels for the audience and to enable interaction. It allows interaction based on real-time multimedia content messaging. It also allows the content creator to send and personalise multimedia messages — audio, video, image, text — to any device connected to the Internet via the mobile app. This allows venues to engage with audiences and get instant feedback from both inside and outside the venue, filtered by the preferences or needs of each end-user in the audience, creating a unique experience.
MyAcc has the potential to go beyond offering accessibility services and can be developed further to obtain data from end-users sitting in the auditorium. Gathering relevant feedback from end-users and adapting the services to target other audience needs in an inclusive approach may pave the way for the creation of new services.
MyAcc has been conceived under the principles of Design for All, looking at accessibility as alternative communication in a given environment. MyAcc aims at rethinking the accessibility chain and integrating all accessibility services in a single platform. It will offer the venues a new channel, allowing better communication and a better understanding of the diverse needs of their audiences, in a more inclusive and engaging way. It will provide the users with a new tool with which to enjoy cultural media content.
This work has been partially funded by the H2020 EasyTV H2020-ICT-2016-2 761999, TRACTION H2020-SC6-TRANSFORMATIONS-870610, ERASMUS+ LTA 2018-1-DE01-KA203-004218, ERASMUS+ EASIT 2018-1-ES01-KA203-05275, RAD PGC2018-096566-B-I00, IMPACT 2019-1-FR01-KA204-062381 and 2019SGR113.
Estella Oncins holds a Phd in Accessibility and Ambiental Intelligence from Universitat Autònoma de Barcelona, Spain. She has a large experience in providing accessibility for live events as a freelance translator, subtitler, surtitler, respeaker for different Spanish televisions and conferences, and as an audiodescriber for Liceu Opera House. Her research areas are audiovisual translation, media accessibility and creative industry. She is currently involved in the Education and Outreach Working Group (EOWG) from W3C. She is a partner in KA2 LTA and IMPACT. She is also a partner in H2020: HELIOS and TRACTION.
Email: estella.oncins@uab.cat
Pilar Orero PhD (UMIST) teaches in the Faculty of Political Science at Universitat Autònoma de Barcelona, Spain. She is a member of the research group TransMedia Catalonia. She is the leader of numerous research projects funded by the Spanish and Catalan Governments related to media accessibility. She participates in UN agency ITU Intersector Rapporteur Group Audiovisual Media Accessibility IRG AVA and ITU-D creating an online course on Media Accessibility. She participates in ISO on standards related to media accessibility. She holds the INDRA Accessible Technologies Chair since 2012 and leads the EU projects HBB4ALL (2013-2016), KA2 ACT (2015-2018) and is a partner in KA2 ADLAB PRO, EASIT, LTA and IMPACT. She is also a partner in H2020: Easy TV, ImAc, HELIOS, REBUILD, TRACTION and SOCLOSE. Website: http://gent.uab.cat/pilarorero/
Email: pilar.orero@uab.cat
Note 1:
HBB4ALL, Hybrid Broadcast Broadband for All, was a EU-funded research project (2013-2016). The project aimed at making cross-platform production and distribution of accessibility features more cost-efficient and yet more flexible to use, and also easier to use.
Return to this point in the text
Note 2:
DTV4ALL, Digital Television for All, was a EU-funded research project (2008-2011). The project aimed at helping to provide accessible digital television programs across the European Union. https://cordis.europa.eu/project/id/224994 (consulted 19.04.2020)
Return to this point in the text
Note 3:
ADLAB, Audio description: Lifelong Access for the Blind (2011-2014) and ADLAB PRO Audio description: A Laboratory for the development of a new professional profile (2016-2019) were EU-funded Erasmus+ projects. ADLAB PRO was the natural evolution of the ADLAB project and aimed at identifying the inconsistencies in AD crafting methods and provision policies at European level and to produce the first reliable and consistent European guidelines for the practice of AD. http://www.adlabproject.eu/
(consulted 19.04.2020)
Return to this point in the text
Note 4:
A detailed description of the UAB in numbers for the academic course 2019-2020 is available online at: https://www.uab.cat/web/about-the-uab/the-uab/the-uab-in-figures-1345668682835.html (consulted 19.04.2020)
Return to this point in the text