We ran two focus groups with people with aphasia to better understand the specific challenges they faced when accessing audiovisual media, complementing the survey results, and allowing us to triangulate the responses more meaningfully. The focus groups involved an in-depth discussion and a video critiquing activity.
4.1 Procedure
Both focus groups followed the same procedure. They were held a week apart and lasted about two and a half hours each. The sessions were video and audio recorded, with participants being given three options on how they wanted to be presented in the final stored videos and video stills: not to be shown at all, to be shown with their face blurred, to be fully visible. The sessions were divided into three main sections: informed consent and demographics, an open discussion of audiovisual media viewing experiences, and a video viewing and critiquing activity, as can be seen in the bottom left image of Figure
3. The session was introduced by a researcher, explaining what this research was about and going through the information sheet and consent form. This, along with all other aspects of the sessions, was supported by the use of appropriate communication strategies, SLTs, and tangible communication aids [
91]. These communication aids consisted of paper-based visual aids, such as a tangible Likert scale participants could point to, and pen and paper, to allow participants to use an additional non-verbal communication method. Following the informed consent, participants filled out a short questionnaire asking demographic questions, such as age and gender identity, their self-perceived language abilities, and questions about their viewing context, such as the devices used, and the type of content consumed.
We started with an icebreaker asking participants to discuss their favourite films. This was followed by an open discussion of barriers participants face when accessing audiovisual media, as well as what facilitates their viewing and what methods they use to overcome the barriers. Participants then engaged in a video viewing and critiquing activity, allowing them to reflect on various aspects of the video clips. We selected 11 video clips that represent a wide range of different types of content, including various broadcast formats (e.g., films, documentaries, news broadcasts) and genres (see Table
3), and different levels of audiovisual media complexity, using a complexity heuristics, such as intense speech (e.g., multiple people talking) and novelty (e.g., new actions happening or a scene change occurring) [
56]. The video clips were relatively short, lasting between 36 and 96 seconds (average = 63.4, SD = 18.2). These choices, while not exhaustive, allowed us to elicit critical reflection to a wide range of common viewing situations participants face. Clips were played on a large screen and projected in front of all participants for the video viewing (see Figure
3) and were introduced prior to their viewing by a researcher. Once the video clip was finished, ample time was given to discuss any aspects of the video clip that participants found challenging, along with characteristics that they found facilitated their viewing. The discussion was assisted by the SLTs who made sure every participant could express their thoughts by asking probing questions (e.g., asking a quiet participant if they found a video clip challenging) and supporting their responses.
4.4 Results of Focus Group Sessions for Reflecting on Challenges and Critiquing Audiovisual Media
We now present the qualitative results from the focus group session, in which participants discussed the barriers they faced, as well as the comments from the speech and language therapists, synthesising the findings from both the reflections from everyday life and the critique of the clips. The transcript for these sessions had 185 references across 26 base codes. From the thematic analysis of the discussions, four main themes emerged:
Understanding Speech,
Cognitive Load,
On-Screen Text and Subtitles, and
Following Narrative (see Figure
5).
4.4.1 Understanding Speech.
The most common issue discussed by participants in the focus group sessions was understanding what people were saying, with numerous characteristics of the speech making it less accessible to people with aphasia. Multiple participants expressed difficulty understanding speech when the speaker did not speak with high clarity. During the open discussion section, P9 talked about a recent experience he had watching a new TV show: “I was watching the Silo uhh yeah and it’s set in a futuristic environment and umm conspiracies are rife and umm that uhh you talk quite lowly umm whisper and I can’t hear it”. This feeling was shared by other participants, with P6 suggesting increasing the sound volume to facilitate listening: “P6: You? [gesturing with a finger, starting from the table moving up, while making a noise that increased in pitch]”. Additionally, P9 mentioned wearing headphones to facilitate viewing, which he could only do when alone since watching socially with others introduced an additional challenge: “R2: But then your wife wouldn’t be able to hear? P9: [laughing] Yeah and even more importantly umm I can’t hear my wife!”. Another participant found it challenging to understand when men spoke, stating “P4: After my stroke, I could not understand men at all, because they not talking like that [gesturing with hand in front of her mouth, opening up from closed fist], they talking so so [mumbling, hand over mouth, head tilted downwards]”. Similarly, many participants found it difficult to understand speech when the speaker had a strong accent or dialect, such as “P9: The Irish accent is... P8: Harder! P9: Hard, yeah... Yeah it is”.
Participants found the fast pace of speech to be challenging: “P1: Because it’s too fast sometimes, so I don’t... so I lose a lot of the words... okay?”. Additionally, participants missed important information when the speech was too fast: “P1: Yeah, there is quiz shows, the Chase? [...] They are so quick, I just glance at it and I thought phew [hand moves over her head]” or “P5: It’s quick, I didn’t register the joke”. Participants stated that once they lose track of what is being said it is hard to continue watching, as well as generally being tiring to keep up. During the video viewing and critiquing activity, certain video clips were deemed to be easier to understand than others because of the pace of narration or speech. One such example was VC1, with participants saying the narration by David Attenborough (famous for his ‘clear’ ‘Received Pronunciation’ English and slow speaking rate) was clear and slow, making it easier to follow: “P3: I find that he’s very good, how he talks, he’s very umm proper, with every word he’s saying”. The clear and simple language used in the narration also facilitated understanding, with the narrator pausing between sentences, allowing time to process the information. On the other hand, VC11 was challenging since both the host and the participants had to speak quickly due to the limited amount of time they had: “P4: I don’t understand her [the host]. She’s always going fast, isn’t she?”. This meant that most participants could not watch such fast-paced quiz shows, even if they enjoyed them: “P5: I like watching quiz programmes [...] It was uhh it was a challenge because it was so quick”.
Participants also found that distracting background sounds could add barriers to their understanding as they prevented the participants from clearly hearing what was being said, including background sound effects: “P9: Road noise, I was aware of the noise... Road noise... P8: Yes, yes. R2: Did it make it harder... what he was saying? P9: Yes it did, because you could hear the uhh rumble”. Additionally, loud background music could also add barriers to understanding: “P7: No, no, not me but uhh for me, no because it is the uhh the music [...] yes, yes. Hard work though because music”. These barriers were exacerbated in video clips that contained both distracting sound effects and loud background music, with some participants completely losing attention: “P8: Noise, speaking, forget it [...] yes, the noise, pops, forget it”.
4.4.2 Cognitive Load.
The second sub-theme involved barriers with cognitive load and processing, a challenge that has been explored in other contexts for people with CCNs, such as people with aphasia during conversation [
12] or people with dementia when playing games [
79]. Participants stated that consuming audiovisual media was often tiring, they found it difficult to keep paying attention if the piece of content introduced certain barriers, with
P5 describing as “
Every time I look at the umm video or watch the television, I have to work hard, much more hard than I used to”. One such barrier involved situations in which multiple people are speaking in a group setting: “
P4: I’m understanding and then suddenly when there is more people on the TV then there is more people and... and I a bit lose it, I don’t really understand it”. Participants found that having one person speaking at a time with clear delineation between speakers facilitated viewing, finding the group conversation in VC4 to be well structured: “
P10: Yes, yes, slow down fine [gesturing with his hand]R2: It was turn by turn, so it was... P10: Yes, good”. Moreover, such temporal delimitation of information was also important for individual speakers, with participants stating that they lost focus when speakers used long and complicated language without having breaks between key points. For instance, when reflecting on VC11,
P1 mentioned that “
I think the- the questions are so wordy, and I lose... [shakes her head, waves her hand away]”. This was also the case for constant streams of information, such as in VC2: “
P10: Voices, umm pausing or long sentences... SLT1: Did [the presenter] pause? P8: Oh no way! [...]SLT1: There was no break? P8, P10: [shake their heads]”.
The cognitive load challenges experienced by people with aphasia make it particularly difficult to concentrate on continuous streams of information from different sources, often resulting in confusion and loss of focus. Participants mentioned the fast pace of constant dialogue or narration, without any breaks, was tiring to follow. Some participants would lose focus and stop paying attention entirely if they did not have enough time to process what was being said by a speaker before the next speaker started: “P5: I listen but then I have to comprehend... umm it takes a minute to comprehend, and then it’s rushing on to the next one and it’s [waves her hand away]”. P2 described trying to keep watching when no pauses are offered to the viewer as “you can’t carry on [waves her hand to go on] because they... you get... so behind”. Similarly, constant fast-paced on-screen action with no visual breaks also introduced access barriers. An example of this occurred when viewing VC9: “P9: The cars were moving too fast... umm I have problems umm concentrating on... the speed the cars are moving, yeah”. When discussing these two barriers, however, participants mentioned that having control over the pacing of the video facilitated viewing, allowing them to either pause the content when they felt overwhelmed, or to rewind it to watch a segment again. Controlling the pace as a coping mechanism would be further improved when watching socially, allowing a co-viewer to help fill in the gaps: “P4: My husband has to stop and explain to me about this and this and this”. In general, the participants repeatedly talked about the pace being a constant barrier, and that slowing down both the dialogue and the action would facilitate viewing significantly.
4.4.3 On-Screen Text and Subtitles.
On-screen text often introduced challenges, including both text present in the shot (e.g., a book the actor is reading) and text superimposed on the shot (e.g., subtitles or news ticker). A common barrier expressed by both the focus group participants and the survey respondents was not having enough time to read the text before it disappears: “P2: Yes but sometimes also the processing... you get it... and then you [moving hand as if reading from the screen]... three-quarters of the way and then changes, you know what I mean?”. This frustration, along with the cognitive demand of focusing on multiple information sources at the same time, meant that some participants did not even try to read the text. Indeed, throughout the focus group session, P9 repeatedly mentioned that he does not try to read any of the text, both in the opening discussion: “I cannot concentrate on the picture and the rolling text”; and when discussing the video clips: “R2: You just still didn’t bother [reading the text]P9: Uh... still concentrating on the speech [...]SLT1: So you’re focusing on the auditory input P9: On the auditory input, yes”. Even when participants were able to keep up with the subtitles, they often missed other important visual information: “P6: Umm... you [imitates reading the subtitles on the screen] Ah! [imitates looking up at the rest of the "image"] Ah, fucking hell! [looks around, then back down at the "subtitles"]SLT1: Yeah, you can’t quite keep up with the text and the picture P6: Yeah [points at "subtitles", looks up at "image", looks surprised]”.
Large amounts of simultaneous on-screen text presented overwhelming amounts of information and introduced additional access barriers. This is different from not having enough time to read it, as the challenging aspect is the amount of information presented to the viewer at once. For instance, when viewing VC2, in which some textual information relevant to the news story was presented alongside the journalist, P4 commented that “Well putting those words together... those words together I... I don’t understand, see? I understand the words, I can say those words, but I can’t put those words together [hand gesture and facial expression of confusion]”. Participants mentioned dealing with this issue by controlling the pace of the content, either by pausing or rewinding. This was deemed to be more useful for on-screen text rather than subtitles, as it gives the viewer ample time to read important information at their own pace: “P4: So sometimes at the end of a movie they will put three pages of writing, and switching umm real films... you know what I mean, um True Love, so [her partner] has to stop, stop it, and then I read it [imitating reading with hand] and then I say go on, and then he goes on and I stop it again and read it little bit”. This is not always possible, however, such as when watching socially or on certain devices: “P9: In order to um for me to watch something with subtitles I have to use my computer laptop and you can control the speed of the... P10: Fantastic! SLT3: Which you can’t do on the TV? P9: No, no, no”. Moreover, participants mentioned that the way this text is presented makes a difference, finding that a clear and segmented presentation of information with simple language facilitated their understanding. Having textual cues, including subtitles, offered the participants additional information that helped them follow along when they felt lost, even without reading all the text: “P9: Sometimes umm it can be uh more accessible kind of in a reinforcement way, because you can scan the text and check to see if you are reading wrong R2: Ah, so if somebody is says something and you don’t understand, then you look at the... P9: Yeah, yeah, yeah”. Similarly, having a second screen with additional textual information also facilitated viewing, especially when paired with controlling the pace: “P2: But then I would go to the BBC news on my tablet, and then I can read it little bit by little bit by little bit”.
4.4.4 Following Narrative.
The final sub-theme deals with barriers around following the narrative of audiovisual media, which participants discussed as a point of frustration throughout the sessions, stemming from multiple accessibility barriers. As previously discussed, the participants mentioned using additional visual and auditory information to piece together the narrative if they lost track of it. When these cues were missing, however, understanding what is happening can become challenging: “P5: You don’t have the uhh actions and uhh visual stimulus, and uhh I couldn’t follow it”. Additionally, visual cues could be presented too quickly for the participants to process them. Repeating the cues, or other information more broadly, helped the participants with understanding: “P8: One time, okay... [rocks his hand side to side, indicating ’so-so’] Two times, fine [waves his hand down]”. Additionally, having clearly segmented pieces of information helped structure the narrative, giving enough time to process the information before continuing: “P7: Yeah, it’s uh short- short this uh [showing small distance with fingers] clip? SLT1: Short questions? P7: Yeah, but it’s [holds up 5 fingers] 5 seconds and stop and 5 seconds... it’s pretty good the whole lot, I can’t R1: Like they had pauses, someone spoke then there was a pause... P7: Yeah, yeah, much better for me, the whole lot no I can’t do it, but this... [n.d.]”.
Sudden narrative or visual shifts were also deemed to be access barriers, with several participants mentioning that they struggled to follow along, such as when viewing VC11: “P1: But, it’s um out of the blue [gesturing, hand waving off], umm so umm the questions umm and then you answer it and I’ve [chuckles, makes a facial expression of confusion] yeah, I can’t”. This was exacerbated by the lack of visual or audio cues that could facilitate their viewing. A similar challenging situation revolved around understanding comedy, with participants mentioning not realising a joke had been made or not understanding the humour in a scene because of the increased effort required of them to understand the dialogue: “P1: I don’t have... don’t have jokes, can’t... follow, I don’t understand it, if you have that problem, yeah? [...] Yeah, no just goes [gesturing her hand flying over her head]”. Understanding humour could be facilitated by audio or visual cues, such as laugh tracks, as they suggest to the viewer that something funny just happened, such as in VC3: “SLT1: What helped you to understand the joke? P9: The laughter SLT1: So the background laughter helped you? P9: Yes, yes”. Complicated narrative structures introduced additional access barriers because of the cognitive effort required to keep track of events: “P9: Back in time you... some stories re-relay their story to... what happens yesterday or future, and you find it confusing”.