Although we employ the three graphicacy stages in both the results and discussion, this is done to provide structure, not to imply that constructing ones understanding of an HDV is a linear process. The three stages are used to describe the activities involved to accomplish each stage’s specific goal and potential barriers to those activities. Rather the process of constructing ones understanding of an HDV is a combination of bottom-up and top-down processes that occur during the interplay between the external HDV and the individual’s interpretation of the information as viewers construct and revise their mental model of the HDV over time.
Additionally, the authors would like to note that the downward cascade to participant performance throughout the various graph reading activities indicated the negative impact of HDVs that did not provide adequate support for people with Down Syndrome as they read and attempted to make sense of the information presented. In other words, a “low” performance does not necessarily indicate the limits of people with Down Syndrome’s abilities; rather it demonstrates how shortcomings in HDV design can introduce unnecessary barriers to effective inference-making, understanding, and engagement with health information.
4.1 Reading Health Data
As described in the background, reading health data visualizations requires sufficient competencies with data, information, numeracy, and print literacies. These literacy-related skills are requisite to notice what is salient, identify the various graph component elements, and recognize if there is any missing information that is necessary to read the HDV. Below we describe people with Down Syndrome’s saliency and sense-making abilities, their identification skills, and some accessibility barriers that can arise during the first stage of reading health data visualizations.
4.1.1 Saliency.
The foundation of reading any graph or chart begins with identifying the various elements that make up the health data visualization. We began our interviews by asking participants some open-end saliency questions, such as “What is the first thing you see?” During the second saliency question, the researcher physically covered their own eyes with their hands and said “I am closing my eyes now so I can’t see.” The researcher then asked participants to describe everything they were seeing in the graph to them. The benefit of taking a saliency-first approach in the interview was two-fold. First, the saliency questions show us generally where the participant was looking and in what order they read the graph elements (i.e., how they naturally read graphs without any kind of structure). Secondly, the act of verbalizing provided insight into the saliency judgments they were making (e.g., an element’s relevance, importance, noticeability, etc.). This combination demonstrated how people with Down Syndrome initially read health graphs without any supports or guidance.
Participants described the HDVs in four different ways (see Appendix
A.3 for all saliency HDVs). First, they made
generalizations based on their initial assessments of the information presented, compared different types of data or regions in the HDV, categorized types of data, and made judgments about the HDV. For example, upon seeing the different images of food in the scatterplot, most participants generalized that the data was split into two groups, healthy and unhealthy foods and beverages. Participants also called out
specific types of graph elements. These could include the type of visualization (i.e., bar or line graph), the use of icons or images, the presence of words, numbers or dates (e.g., “I see a lot of numbers” in the table). Participants also vocalized
elements’ specific text content, such as the title in verbatim or an actual number value (e.g., “2.0”) in the HDV. The last type of observation that our participants made was noting the various
descriptive qualities (e.g., the color, size, or position) of HDV elements.
All participants made multiple combinations of descriptions for every HDV. This demonstrated that every participant was capable of varying levels of abstract thinking upon seeing an HDV for the first time. The number of observations verbalized also differed between participants. The quantity and locations of observations similarly indicate different graph reading patterns across our participants. It also showed the visual path they were taking, where they visually focused on a region or if they returned to a region or element more than once. As such, the order of the observations likewise suggests varying levels of ability in effective scanning of information.
When initially reviewing the HDV, participants often leaned upon skills they were strongest in. For example, most participants employed their print literacy skills first. Many participants demonstrated a tendency to read specific text elements first. Participants then read text usually from the largest to smallest in text size. This was followed by participants verbalizing items that had sufficient color contrast as this facilitated quick and easy recognition of information. After the written text, participants demonstrated a tendency to either call out familiar icons or images or specific colors when present in the visualization. Next, participants verbalized numerical information. The order of numerical elements that were verbalized similarly followed the largest to smallest with strongest to weakest color contrast. This suggests that familiarity and confidence in various skills may influence what is most salient and in what order this population reads HDVs. In other words, many individuals were immediately drawn to HDV elements they were confident in their ability to make sense of the more familiar elements (e.g., print, numbers). This familiarity-first behavior suggests some viewers with Down Syndrome may employ a combination of top-down (i.e., long-term knowledge and skills) and bottom-up (i.e., taking in stimuli without context) information processing as they interpreted the information to construct their understanding of an HDV at the earliest stage of graph interaction.
As people with Down Syndrome can struggle to express themselves, it is worth noting that several participants frequently avoid saying words or numbers that they struggled to audibly say. For example, many participants struggled with saying the word “macronutrient” in the stacked bar chart. Instead, they talked around, were hesitant to say, or omitted the word (e.g., “Whatever that long word is called” [Harper], “It’s a bit hard to say” [Morgan] “This week’s–not–uh, I–I don’t know” [Jordan]). However, participants struggled less when they broke down the syllables. When Emery got to graph #3, their study partner covered up the syllables as they read them, allowing their to gradually read the word. This may signal an accessible design opportunity for long words or jargon in HDVs with populations who may struggle to visually parse through multisyllabic words (e.g., “This week’s Mac-ro-nu-tri-ents”).
Similarly, verbalizing numerical information also highlighted the numeracy issues people with Down Syndrome can experience. There appeared to be an issue when reading long number place values. Past work has suggested that dyscalculia is often a part of the Down Syndrome behavioral phenotype [
30], which occurred when participants encountered extra characters, such as decimals separating the whole and fractional numbers, commas indicating higher-level number place values (i.e., tens, hundreds, thousands), or dashes used in ranges. For example, Emery read the age range “31–35” as “135” in the first table. When Jesse said, “18 and 20, and 20 and 39, and 62, 85,” they were actually reading the age range (18–20), the first two cells (2.0, 3.9) in the blue Lean region and first two cells (6.2, 8.5) in the green Ideal table region. Similarly, Cameron read the total value of 63,451 in the bar chart as “Six thousand–six, five, thirty-four, fifty-one.” Skyler had to correct themselves when reading the average number of steps “9,600–9,064.” This may indicate incorrect number articulation errors caused by special characters creating visual shifts as they read the number and encoded the position with the appropriate number place values.
Invisible number lines could also be a stumbling block for some participants. Jesse similarly could not find their age: “I’m 22. but 22’s not on here.” According to the table, Jesse would fall into the 21–25 age bracket. In this example, the use of age ranges requires a viewer to have both sufficient working memory and numeracy skills to recognize the invisible number line with a range of numbers. Increments on the X- and Y-Axes may also introduce invisible number line barriers. On the “Weekly Walk Distance History” line graph, Darcy observed: “I see that the graph skipped some numbers. It starts from 2.0 and then it is counting by five. Five, zero, five.” Sloane struggled to articulate their frustration with the increments–in the y-axis of the macronutrients stacked bar chart. They referred to these skipped numbers as categories: “I see 100 category. I see two–2,000 in categories and three cat–category and forty catty–category, ach!”
4.1.2 Identifying Health Data Visualization Elements.
An essential part of reading a health data visualization is the ability to identify the various graph component elements (i.e., title, axes labels and values, icons, images). HDV element identification requires sense-making, spatial awareness, and effective scanning abilities. Additionally, interaction with a data visualization is not a linear process [
17]. Instead, someone’s interaction and understanding of an HDV is continually refined as more information is visually sensed and encoded and mental models for the data are iteratively revised and updated. Table
2 reports upon participant performance of each sub-task within this first stage of HDV reading. Correct answers are indicated by a point and incorrect answers with zero points for each graph and each corresponding question.
Overall, the participants performed well when identifying graph elements such as the: title, X-axis labels and values, Y-axis labels and values, values within the visualization, and various symbols, icons, or images. They also were able to identify the various stylistic elements such as shape, color, and size. Six of the ten participants were able to identify more than 75% of the various HDV component parts. The remaining four were able to identify between 65.5% to 72.4% of the elements. This suggests that the graph perceptual, sense-making skills of people with Down Syndrome are relatively strong during early health data visualization reading identification activities. This is in line with past work that found people with Down Syndrome to be strong visual learners [
41].
4.1.2.1 The X- and Y-Axes. The X- and Y-axes labels scored the lowest during graph element identification. This appeared to occur, in part, because of a behavioral reaction to pictures when influenced strongly by existing health knowledge particularly in the scatterplot graph (graph #6). The scatter plot’s use of strong, realistic pictures appeared to reinforce our participant’s existing nutritional understanding. As a result, the X- and Y-axes labels were often ignored. This visual disregard may have occurred because photographs of food were used. Viewers appeared to instead fixate upon the highly familiar, concrete data points rather than noticing what was being compared on the scatter plot, which compared the percentage of Americans versus Nutritionists who said whether a food or drink was healthy or not.
Participants demonstrated an observable tendency to first notice the X-axis, followed by the Y-axis on the left whenever participants read graphs. However, many participants (60%) struggled to notice an additional Y-axis was included in graph 5. Only three participants noticed both the label and the percentage values [Harper, Skyler, and Sloane] when a dual Y-axis was present. This behavior may indicate that some participants were unfamiliar with the procedure when reading different types of graphs with axes in different quantities and locations.
4.1.2.2 Icons vs. Images. In the HDVs that used icons (graph # 2–5), participants verbalized more descriptive qualities (e.g., colors, or data types, such as numbers or words) when they generalized what they were seeing in addition to a greater number of specific HDV details as they took in the information. Conversely, most participants (80%) in the scatter plot HDV used more generalized descriptions of large categories of information. It appeared that the use of images of various food and drinks combined with an unfamiliar graph type caused them to rely more upon their understanding of nutrition. This resulted in broad generalizations about the data points and the visualization as a whole that was directly informed by their pre-existing nutritional knowledge. Participants also categorized and grouped the more familiar imagery (e.g., foods and drinks, healthy and unhealthy, “fat stuff” and “too much salty” [Morgan]). This kind of data categorization and grouping can support estimation abilities [
119]. Participants, like Cameron, made other associations with what they were seeing as well: “I see different types of ... foods in the kitchen, oven and stuff and the cups.”
Most participants (80%) skipped a lot of individual data point identification that they had demonstrated in previous HDVs. Instead, many went straight to making judgments about what was being depicted. This interestingly suggests that mental models of HDVs could become less flexible if pictures are used on their own without additional elements that can support accurate inference-making if an assumption is made based solely on the familiarity of the images in conjunction with their initial impressions of the data. It appeared that the confidence experienced by the participants caused by the combination of familiar presentation and topic that simultaneously reinforced their existing nutritional knowledge made them less visually critical of the remaining HDV elements and what the relationship between data points could mean, especially if it conflicted with their existing understanding of the topic being visualized.
4.2 Reading Between Health Data
Reading between health data in visualizations requires even more skills (e.g., health data literacy and numeracy skills, print literacy, abstract and spatial reasoning, and ratio processing abilities) to effectively interpret visualized health information. These skills are critical for HDV viewers to: (1) encode and map the information and (2) compare values. This section details the inference-making skills of people with Down Syndrome as they encode, map, connect, compare the visualized health data.
Table
3 reports upon participant performance of each sub-task within the second stage of HDV reading. During the various graph reading connection activities, performance diminished (32.42%) among our participants as compared to identification activities. Only two individuals were able to satisfactorily answer more than 75% of the questions in this stage. Three fell into the third quarter (50–75%) and the other half of the participants scored 50% or less. These results point to potential accessibility and HDV design opportunities that better support people with Down Syndrome as they connect, compare, and interpret graphs. Please note: partial points (i.e., .5) were awarded when participant answers were close, but not entirely correct.
4.2.1 Encoding and Mapping Information.
Mapping information in HDVs generally consists of connecting the identified component elements of the graph’s anatomy to each other and encoding the meaning of visual attributes. Mapping information in this way supports viewers’ ability to assign meaning to each connected element and update their overall understanding of the visualization. In this subsection, we describe the two visual attributes these HDV employed to support interpretation of categorical information encoding: color and images.
4.2.1.1 Color Encoding and Meaning Mapping. Color is typically used to distinguish categorical information by grouping elements so viewers can more easily identify similarities and differences. Many systems designed for people with Down Syndrome or other IDDs, use color-coding to indicate more than just groups. Some nutrition-oriented health apps use the color metaphor of a stop light to indicate a food’s health status (e.g., [
71,
101]. However, graphical properties, like color, are not equal in their ability to accurately communicate meaning. Other channels, such as spatial region, position, length, angle, and size, are more effective [
25,
84].
Three graphs used color encoding: the table (#1), the stacked bar (#3), and the line graph with two y-axes (#5). The table used both colors and labels to associate the two as a group (i.e., blue = lean, green = ideal, yellow = average, red = above average body fat percentage). In the line graph, color coding was used to indicate intensity of a physical activity to visually link the heart rate and effort y-axes together. Finally, the stacked bar used color to indicate the healthiness of a macronutrient using the stoplight visual metaphor, to distinguish between the three values, and to visually link the macronutrient label with the icon (i.e., lettuce = healthy carbs, leg of meat = protein, butter = fats).
When color-coding did occur, 30% of the participants had no meaning association. Instead, viewers, like Jesse, inferred that color was simply a stylistic choice: “it means the colors ... like different kinds of colors. It’s blue, green, yellow, red.” The percentage of those who had no meaning associated with color was higher during the first half of the interview versus after spending more time engaging with the HDV. Harper was the only one who explicitly stated that the graphs were “color-coded.”
In the table, many participants recalled a pre-existing color association, “Normally, the green, yellow, and red means ... Stop, Slow, Go. But [I’m] not sure about the blue” [Harper]. When other colors did not also map to the metaphor, like the blue, confusion occurred. For Jordan, green was “good,” yellow was “bad,” and red was “very bad.” However, they instead associated blue with the affective state of “sad,” which may indicate a color metaphor-mismatch occurred. This inaccurate encoding made interpreting the graph more difficult. Typically a high-performer, Cameron did not notice the labels at all. Instead they interpreted the color as corresponding with the size of the region. While Sloane did connect blue with its lean label, they said green was “healthy” and red was “really bad.” They described yellow in terms of foods that were both healthy and yellow-colored.
Graph #5 similarly indicated how more explicit mapping between color and meaning is necessary in more accessible HDVs viewed by people with Down Syndrome. High-performers were able to associate how color was used. Skyler observed that colors were the “different kind of colors of different beats in your heart.” Similarly, Cameron was able to connect the title and associate the color with its gradation: “The colors mean how–how deep is the intensity... [Green] means like not that–not that intense. Dark red means that it’s that extreme amount in the intensity.” However, imperfect color encoding and mapping still occurred 45% of the time. Jordan associated the color with the icons rather than intensity levels. This suggests that despite the presence of labels, color is not strong enough to ensure accurate mental connections in a world of potential implied meanings.
The stacked bar chart had the highest level of correct color mapping at 95% accuracy. Even participants, who consistently struggled [Shiloh, Emery, and Harper], were able to connect both color and icon encoding when properly reinforced with familiar distinct imagery that was reinforced by understanding of health information and had a label to support mapping. However Jordan’s concrete associations with the food icons overruled a totally accurate color interpretation: “[Red is] bad food. [Yellow is] good food. [And green is my] favorite food.” This may signal the strength of lived experience to inform HDV interpretation.
4.2.1.2 Image Encoding and Meaning Mapping. Five of the six HDVs used icons or pictures to support comprehension as suggested by previous accessible visualization recommendations [
134]. While icons did support most participants’ connection between graph elements, some participants did not like the use of icons. Darcy, who self-reported that they were very familiar with reading graphs, preferred a “dot” instead of the shoe icon. As such, there are some caveats for how those icons could be incorporated into visualizations intended to be more accessible to people with Down Syndrome within the health context.
Appropriate image selection is critical. Sloane initially thought the vertically stacked shoe icons in the Daily Steps bar graph was a “shoe store.” While Cameron extrapolated that “those lo–logos represents [the] amount of di–distance” in the Weekly Walk Distance line graph. Others, like Jordan, associated the shoe icons not with steps but with physical activity (i.e., “walking”) in the bar graph. However, for Darcy, the shoes were simply shoes and the caution sign meant nothing to Morgan in the dual Y-axis line graph. Sloane interpreted the lettuce icon in the stacked bar chart as green brain, which, in turn, impacted their connection of the icon to its healthy carbohydrates label.
Graph complexity can also interfere with accurate icon encoding. The use of multiple icons in a single graph should be carefully considered, particularly if there is not a label to provide a redundant encoding for bundling connections between elements. In the fifth graph, while the use of different icons were intended to visually represent a change in activity intensity to reinforce position on the graph as well as the color encoding on the graph background, the different icons did not clearly map. As a result, the complexity and use of multiple icons in the dual y-axis line graph contributed to the icons only being 45% accurately encoded–the lowest scoring graph.
Icon encoding can take time to process. Like some, Morgan did not initially verbally associate the shoe icons with steps. However after spending time answering questions about the bar graph, those participants persistently associated the shoe icons with steps across all three graphs that used the shoe icon. This was interesting as the meaning of the shoe icon changed with every graph that used it (i.e., bar graph #2 = steps; line graph #4 = distance; dual y-axis line graph #5 = mid-level intensity). The re-use of icons when an association has been made caused participants, like Morgan and Harper, to consistently demonstrate this carryover effect during icon encoding. Other icons also had pre-existing associations, which similarly affected correct meaning-mapping. The use of hearts and a warning sign to indicate higher heart rate zones meant something different for Harper, Morgan, and Jordan, who interpreted the hearts as “love” or “falling in love” and the triangular warning symbol as “person” [Harper], “danger” [Shiloh and Sloane] or a “danger zone” [Darcy].
The results from this and the previous sections may signal an accessible HDV design feature opportunity to provide more explicit inference-making features that support the connection-making abilities of individuals connecting graph elements and channel encoding, like color or icons.
4.2.2 Comparing Information.
Comparing information in HDVs involves both visuo-spatial and abstract reasoning as well as ratio-processing abilities to recognize patterns, similarities, differences, extremes, anomalies, and ultimately connect the various elements together to infer the overall trend. This section describes how people with Down Syndrome compared high and low extreme values. It also reports upon the differences participants noticed when viewing HDVs. The number and kind of differences indicated varying levels of cognitive flexibility among our sample when recognizing patterns, generalizing, and grouping.
4.2.2.1 Comparing Extreme Values. The extreme value questions illustrated authentic reading tasks where the viewer begins with a specific question to find an exact value (see Tables
14 and
15 in Appendix
A.2 for the exact HDV specific questions asked). Overall, participants performed moderately well when comparing data to determine extreme values. They were able to accurately discern both the high values roughly half of the time: Table (60%), Bar (50%), general values Stacked Bar (50%) specific values Stacked Bar (60%), Line (45%), Dual Y Line (45%), and scatter plot (45%). They performed slightly better when judging low values: Table (85%), Bar (70%), general values Stacked Bar (50%), specific values Stacked Bar (60%), Line (50%), Dual Y Line (55%), and scatter plot (40%). One reason for the moderate performance may be that participants had to use different procedural skills to locate extreme values across all six HDVs.
For example, the table (HDV #1) required effective scanning of a large amount of numbers, some with poor color contrast. While Shiloh generalized to entire regions, some participants, such as Morgan and Sloane, instead visually fixated and answered within their field of vision when the question was asked. The table also required participants to recognize patterns across the overall numbers to find the location of extremes. The values increased from top-to-bottom and left-to-right with the lowest in the upper-left and the highest in the bottom-right. There were also table reading procedural issues for Jesse and Sloane, who answered with age values, which were literally the highest number visible to them.
HDVs #2-5 required participants to visually track between the graph’s axes and the individual values. During this visual back-and-forth, participants engaged their ratio-processing system to visually compare the differences between values. The brain’s ratio-processing system attends to ratios of difference between non-symbolic values (i.e., not number symbols, but shapes or areas) [
82]. As HDVs are visualizations of non-symbolic values, comparing between them requires noticing fractional differences between the information represented. When the contrast between values is great, it requires less cognitive effort. When it is smaller, it can increase a viewer’s cognitive load (Figure
6). After this comparison has occurred, they then must keep track of each of the value judgments in their visuo-spatial working memory until they find their answer.
Comparing the lowest values in the bar graph required the least cognitive effort for ratio-processing. This made sense as most participants were able to correctly answer this question. However, more sensitive ratio-processing skills were necessary for the high values in the bar graph and both the overall high and low extremes as well as the specific macronutrient type extremes in the stacked bar graph. Half of the participants struggled to correctly determine the highest value in the Daily Steps bar chart. Saturday (i.e., the correct answer), Tuesday, Wednesday, and Thursday all had very similar values to each other. Because of this lower contrast ratio, four out of the five answered one of the visually similar, yet incorrect values (Emery, Harper, Darcy, and Jordan).
Ratio-processing skills were taxed in the stacked bar graph when participants were asked to determine extremes of specific macronutrient types (i.e., highest protein and lowest fat). Unlike comparing values in the bar graph, which started on the same level on the x-axis, comparing fats and protein areas were more difficult because they sat on top of the different healthy carbs values. Being stacked on an uneven base appeared to impact participant’s abilities to effectively process and differentiate between the lower contrast, fractional differences in the sizes of the protein and fat rectangles. This suggests an accessible visualization design opportunity to highlight when the values have low ratio contrast and reduce unnecessary impacts to the viewer’s cognitive load.
When participants had hit their ratio-processing limit, they reverted to a more familiar graph reading strategy of looking to the top of the shape to determine the highest or lowest values. All four of the participants (Emery, Harper, Darcy, and Cameron), who incorrectly answered Thursday as the day with the most fats, did so because Thursday had the most overall grams. Visually, Thursday had the highest position from the top. Similarly, when Shiloh was uncertain how to judge the highest and lowest values in the daily steps bar graph, they reverted back to their stronger number reading skills as they seemed less confident in the ratio-processing abilities. In both instances, Shiloh answered with the smaller numerical metric underneath the title: “Yeah, so the highest and the lowest: 63,451. And the lowest is the nine thousand. No, no. Nine hundred, sixty-four ... so–the highest and the lowest.”
While the imagery was the most familiar, the procedural skills required to read a scatterplot graph was the most foreign to all participants. Several participants leveraged other graphicacy skills they felt more confident using as they interpreted the nutritional data. For example, two participants reported that kale [Darcy and Sloane], which was visually in the highest position to the top of the graph, was the most healthy. Other participants relied upon their health literacy and were instead influenced directly by their nutritional knowledge. In the scatter plot, participants' broad generalizations were impacted by their existing understanding of nutrition: “the highest food is the healthy foods” [Skyler], “the veggies up top” [Emery] or “the fruit and vegetables” [Harper]. Shiloh said the “junk. It’s ice cream–the desserts” were the least healthy in the scatterplot. Harper echoed this assessment by generalizing with “all the junk.” As a result, how participants compared values suggests that when individuals with Down Syndrome felt less confident in their interpretive procedural skills they instead switched to more familiar skill sets and prior knowledge to interpret HDVs. In other words, they use the same tactics as other typically developing populations when they are not sure how best to proceed when they are uncertain how to interpret a data visualization.
4.2.2.2 Comparing Differences. The ability to effectively compare values within an HDV is critical to notice patterns within the data, where values diverge from each other, and what anomalies or outliers there may be. Effective comparative skills support the viewer’s understanding of the data by examining the relationships between values. Comparing the relationships between and across the HDV is foundational to interpreting patterns on the micro (e.g., [in]consistency of a performance instance) and macro level (e.g., overall trend across multiple instances).
All of the participants were able to compare data and demonstrated varying levels of cognitive flexibility when answering this question. Skyler showed the highest level of flexibility when mentioning differences in HDVs–a total of 18 differences across the six graphs. Cameron was the second highest at twelve and Emery reported eight. The lowest was Harper, who noticed three differences.
Participants verbalized multiple kinds of differences. The types of differences mentioned were: the individual shapes, differences between regions, overall trend across the visualizations, colors, and images. The most commonly reported difference type was the various kinds of data (i.e., text, numbers, axis labels and values, specific graph values). This once again highlights our sample’s tendency to rely upon and leverage their strongest skills (i.e., reading) when doing an unfamiliar task like verbalizing differences of data representations.
4.2.3 Connecting Relationships across HDV Elements.
After an individual has encoded and mapped meaning between the elements of the HDV and has compared values to get the gist of the data, the viewer will then connect those component elements together. Connecting information allows viewers to make sense and begin to understand the overall nature of the HDV. Creating mental relationships between the information allows the viewer to understand: (1) the topic and (2) the overall trend.
Although most participants performed moderately well throughout the entire interview, performance dropped dramatically as many participants struggled to connect the various elements and aspects of an HDV to synthesize a coherent understanding of the information represented. Results support previous work which found that while people with Down Syndrome understand abstract information, the differences to their working memory can make managing too much information with too many relationships at the same time a challenge [
20,
60].
4.2.3.1 Identifying the Topic. Only a quarter of the participants were able to connect the various graph elements, aspects, and information to the overall topic of the table (HDV #1). A little less than half (45%) of the participants were able to get the overarching topic of the bar chart, the stacked bar, and the dual Y-axis line graph. The HDVs with the highest levels of connection between the data and the topic were the line graph and the scatterplot at 50%. It is worth noting that unfamiliarity with the scatterplot graph type led to everyone generalizing. Everyone got partial credit for verbalizing that the graph depicted the overall topic of healthy and unhealthy foods and drinks. However, no one was able to provide the more nuanced answer: the scatterplot was comparing the perceptions of healthiness of food and drink items judged by nutritionists versus the average American.
Cameron, Darcy, Emery, and Harper were the most consistent individuals to succinctly synthesize and summarize many of the HDVs. Harper described the table being: “about the ages and the percent of the fat.” Emery connected the HDV to their everyday life, which made connecting the data to the topic much easier: “It is called steps for–same as my watch! It tells you, like, activities in there.” Cameron recognized that the macronutrient stacked bar graph was about: “the grams of, like, the amount of food. And the food has different categories because carbs, protein, and fats.” Emery described the line graph as: “It looks like a snake. That is how many walks have you done. ... how much you’ve done it–of the walk distance history.” Cameron summarized the topic of the dual Y-axis line graph by saying: “it’s all about the intensity in the activity, um, it tells you ... the times at the bottom. Um, is telling you about different times of the levels of the activity intensity.”
A mixture of partially correct and incorrect answers indicated that participants were influenced by their personal understanding of health, exercise, and nutritional knowledge. For example, Skyler said the table was about: “how much you eat ... It has different kind of colors of what–what the–the healthiest things that you can eat. That’s it in my head.” Like Jesse, Darcy described the Body Fat table as: “different pounds of weight you have and also about your losing.” Their response was informed by the visual similarities between the much more familiar BMI chart often seen in doctor’s offices. Emery recalled the food pyramid when they saw the stacked macronutrients bar graph: “it’s ... like, um, a food triangle one that is ... And then there’s a different one. Different one is, like, that equals healthy one, protein, and fat.”
When participants answered incorrectly, most responses consisted of describing and identifying elements and aspects of the HDV rather than connecting everything together. For example, Sloane described the stacked bar as: “combine as healthy. And it will really combine to protein and fat” and counted the number of data points in the dual y-axis line graph, which had “22. There’s 22 times. It’s something measuring from, uh, to 80. Uh, maybe 75. ... all about, uh, the line of, um, gray mark.”
The issue of multiple blocks of easier-to-read text came up during the topic question as well. For some participants, seeing these text blocks did cause some incorrect connections. “It’s graph about the weekend because it started on, uh, every month. Like, um, Jan–January, Feb–February, March, April, and May. Hmm... new year. I don’t know” [Sloane]. “It’s the steps been taken in the years—since the numbers” [Harper].
4.2.3.2 Identifying the Overall Trend. Trend identification was the most difficult question for our participants. Most participants struggled to connect how each of the individual data points worked together to describe the overall trend (i.e., how the parts describe the whole).
Participants answered correctly the most often when the trend was obvious. Participants performed the best in the Weekly Walking Distance History line graph at 40% correct identification of the upwards trend. Several participants actually described the overall trend as they were connecting the graph elements to the topic. “This graph is describing that the walk distance is increasing” [Darcy]. Cameron described the red line as: “a snake goes up. ... all the logos in the snake that–that goes up th–those lo–logos represents amount of d–distance.” When Sloane couldn’t find the words, they instead vocalized the change: “the highlighted [line] And it goes “whoooop.” ... the number [is] bigger.” They audibly changed pitch of the vowels from lower to higher as they said “whoooop” to express the changes in the increasing trend. Sloane again embodied their response to describe the positive trend in the scatterplot as well. Cameron drew upon their health and data literacy skills to demonstrate their understanding of how nutritional components of the foods depicted affected where they fall on the plot: “The nutrients number could change because of the sugar weight.”
Trend identification was particularly challenging in the table and the scatterplot. This is because trend identification in the table required observing the changes using the numerical information only and none of our participants could recall interacting with a scatterplot before, so unfamiliarity of the graph type made describing its trend difficult. It was also challenging when there was no discernible pattern (i.e., not clearly ascending, descending, or remaining around the same amount), which occurred in the bar graph. Of those who were able to correctly identify the trend, only Cameron could describe the trend in the table and the scatterplot. Darcy was the only participant who partially described the bar graph’s trend and two others were able to identify the stacked bar and dual y-axis line graph’s trends.
There appeared to be a relationship between participants who were detailed describers during the saliency questions and those who were able to describe the HDV’s trend. Many participants used more concrete language to describe how the visual changes in data points appeared across the entire visualization. Skyler, who excelled at descriptions throughout their interviews described the stacked bar’s trend as: “I know for a fact it’s wavy.” Oftentimes relating the abstract patterns to more familiar, concrete imagery made trend interpretation easier to articulate. Shiloh similarly described the dual y-axis line graph’s trend: “it’s like a noodle.” In the same HDV, Skyler described the initial upward climb and subsequent dips as “very like up and down. It’s like a roller coaster. ... a pool. But it goes straight, but it has the little roller coaster baby pool to me.”
4.3 Reading Beyond the Data
Data visualization literacy activities associated with reading beyond the health data are critical for viewers to reach the interactive and critical health literacy levels. Much like the interactive level of Nutbeam’s health literacy model (see Section
2.2.4), reading beyond the data requires HDV viewers to engage, ask questions, leverage and apply new health information to decision-making. This is then mapped to examples of applications of those skills as they more deeply engage with the HDVs. This section reports on how (1) people with Down Syndrome would engage with their data, (2) their information-seeking inclination during HDV engagement, and (3) what kinds of changes they thought they should make given the information presented in the graphs. Please Note: Partial points of .5 were given in Table
4 when participant answers were close, but not entirely correct.
Overall, our participants demonstrated low levels of the more advanced engagement stage-specific skills. There was a further 29.24% performance drop from the 2nd stage to the final engagement stage. Participants demonstrated mostly lower levels of engagement with the presented health information. Skyler was the only participant in the final stage whose behavior indicated they was engaging with the HDVs. Only two participants, Emery and Darcy indicated moderate engagement. Shiloh, Jesse, Cameron, and Sloane have low levels of engagement. Harper, Morgan, and Jordan demonstrated very little engagement.
4.3.1 Interaction Potential and Expectations.
A well-known data visualization design mantra in HCI suggests that users want an “overview first, zoom and filter, then details-on-demand” as they engage more deeply with whatever information is visualized [
115]. However, these best practices may not be as intuitive for people with Down Syndrome, who often did not think HDVs were interactive. 60% of participants thought that the HDVs were not something they could click on (table and bar: 75%, dual y-axis and scatterplot: 60%, line: 50% and stacked bar: 30%). Then, we asked what they thought would happen if they did interact with it. More participants thought of potential ideas for interaction (bar, dual y-axis, and scatterplot: 50%, table and line: 40%, stacked: 30%).
Many said they could not mentally envision or had “no clue” [Harper and Shiloh] what would happen if they chose to click on anything. The unknown outcome made even the adventurous computer user, Shiloh, who will “Click click click click.
Gestures clicking across two computers. ... all the time,” irresolute. Instead, unfamiliarity with the task domain appeared to increase their hesitation and reduced their desire to perform even exploratory clicks. Jordan confirmed that any click or interaction with a HDV would be a “surprise” to them as they could not imagine what would happen. When Jesse did finally click on the stacked bar graph, they said: “Uh I broke it.” They immediately blamed themselves–rather than the HDV or technology–for the lack of response from their action. This may indicate those who were similarly hesitant or declined to interact may also share a strong internalized locus of control that could contribute to feelings of low self-efficacy during unfamiliar tasks [
109] like engaging with HDVs.
Only Skyler interacted with every HDV with unprompted, adventurous exploratory interactions. Others just verbalized where they would click or tap. Participants said they expected interactivity on the axis values or labels [Emery, Shiloh], graph values, such as numbers, lines, or bar graphs [Cameron, Jesse, and Jordan], icons or images [Shiloh, Jordan], or “anywhere” in general [Harper, Emery]. While hesitant to click, Shiloh and Morgan did use Google to look up terms they were uncertain about, like “clickable.” Shiloh, in particular, regularly used the browser to find definitions throughout the entire interview.
Of the seven participants who did expect something to occur after they interacted with it, only two mentioned expecting details-on-demand for most of the visualizations [Darcy and Cameron]. They both expected to see the same, additional weight information on the body fat table cells, the total steps for each bar, the “carbs, proteins and fats” [Darcy] “by the amount of the grams” [Cameron] for each day’s macronutrients, both the number of miles and the distance in the line graph, and the nutritional information in the scatterplot. The two only diverged in the dual Y-axis. Darcy wanted each icon to reveal the “activity you’re doing” and Cameron instead expected to see “the number amount of the intensity.”
Some participants also expected the same type of outcome for each kind of interaction with the HDV. Emery wanted clicking on the HDV to either hide or close it except for the bar graph. For the Daily steps, Emery wanted it to function just like the same visualization they regularly interacted with on their Apple smartwatch. “It’s going to put it three ways. The first one is the activity thing. Second one is the serving thing. And then [the] third one is the rewarding, um, award thing.”
However, almost everyone had different expectations for what that interaction should be. Jesse wanted different types of interaction outcomes upon clicking the HDV, such as video content from YouTube in the table, “An app comes up ... Like social media” for both bar graphs and “my FitBit” for the stacked bar, and “a story or ... like a movie or something” in the line graph. Others expected an animation of more information to “... come right at you, ... get bigger” (Skyler) and “pop right up” (Shiloh). Sloane liked the idea of having videos or agents to provide additional information support. “Like a chat ... like, uh, people talking” about how fats can be “really unhealthy ... not healthy at all.” Like Sloane, Skyler also suggested gamified elements: “the shoes will come flying at you. ... I would typically duck under ... and you can eat the lollipops” [as a reward]. Sloane described how icons could use animation, which might provide additional layers of redundant meaning encoding to better support viewer interpretation and conceptually link the image to the activity being visualized. They suggested animated icons like the “walking of the shoe” and “a heart beating” to reinforce the connection with BPM.
4.3.2 Information-Seeking and Question Generation.
Even though participants didn’t generate questions or have information-seeking triggers when interacting with the HDVs around three-quarters of the time, the remaining 25% generated rich types of questions and info-seeking activities. These were oriented around: (1) wanting more explicitly connected information together, (2) wanting additional encoding supports for icon meaning and value, (3) step-by-step guides when they were uncertain how to progress, (4) wanting to limit visual messiness, and (5) clearer definitions for unfamiliar words or abstract concepts. This section also describes the visual aversion we observed in some participants when they encountered intimidating or unfamiliar content in HDVs.
Emery had difficulty connecting the relationships between various types of data. Emery wanted these connections to be “better explained.” Shiloh was likewise uncertain how the age brackets and the regions were related and what those connections meant. Skyler also wanted to support connecting the relationships between various graph elements together (i.e., how the walk distance line graph values and weeks in the x axis were related). Emery confirmed that they wanted additional information and support interpreting values to better understand why the extremes happened and what other values meant in the stacked bar chart. Finally, Sloane wanted an explanation of the trend in the Weekly Walking line graph.
When the icons were used, Skyler wanted to know what the shoes represented numerically. “I wish it was explained better by how many shoes are on the measuring sticks.” If icons are meant to support comprehension, that needs to be explicitly explained in the design. However, if each icon equals a specific number of items, the total rather than an implied calculation needs to be included. In other words, a legend that says “1 shoe = 500 steps” would be less effective than having the total label say “On Monday, you walked 9,400 Steps.”
Finally, several participants wanted the HDVs to be more neatly ordered as visual messiness hindered their ability to make connections. For example, Skyler wondered why the color blocks in the table were different sizes. As oftentimes size is used to indicate quantity, some participants were confused when they noticed that the red above average block was visually larger than the blue lean region, which was, in turn, smaller than the green ideal and yellow average classifications. Emery, Cameron, and Darcy wanted neater categorizations of healthy and unhealthy items as they were visually overwhelmed by the data points being “jammed together” [Cameron] in the scatterplot. Darcy suggested filtering content and drill-down functionality to make it easier to view. “I wish they would organize the graph better. Yeah, because like now it is like messy ... so someone would click on carb, and then all the carbs would go into a category, and then there would be protein it would go there and then fats would go there into different groups–3 groups.”
Quick to look up information in the browser, Shiloh immediately Googled the word “intensity” when they encountered it in the dual-y axis line graph. However, they grew frustrated at times with the unfamiliarity and difficulty of the task domain (i.e., reading, interpreting and engaging with HDVs) when search engines did not provide helpful results. When Jesse encountered the unfamiliar words and concepts of “healthy carbs. [and] mac-uh-nutrients,” they wanted additional information to better understand what these items meant. Jesse also wanted definitions for unfamiliar health metrics: “I don’t know what beats per minute B–BPM means.” Skyler instead preferred to ask people in their life what things meant.
For other participants, however, they demonstrated a visual aversion to unfamiliar graph elements or aspects that felt either unnecessarily difficult or intimidating to parse: “I’m not clicking at the words” [Morgan]. Throughout the interview, many participants appeared to prioritize their cognitive effort and adjusted what they chose to attend to. As a result, more abstract or unknown elements became blind spots to be ignored as the viewer attended to problems they felt confident enough in their skills to solve. Another example of this behavior was when many participants fell back on their print and number reading skills when they were uncertain of the steps required to interpret and understand the abstract visualizations.
4.3.3 Reflecting upon Personal Behavior Changes.
Participants varied in the depth of the level of reflection they were able to engage in with the HDVs. For example, Jordan connected the Body Fat table to being more conscientious of their food intake by “eating salad.” Morgan also found the line graph to be a motivation for them: “It is telling me to walk a little bit far, and like ‘keep going, going, going’ and like ‘take a break to drink lots of water’ [and then] ‘keep walking, walking again.”’ An athlete, Darcy, who already tracked health information in various aspects of their life, often found inspiration for other ways they could leverage their personal information to become even more physically fit with every visualization: “The [table] is telling me I need to start, I need to weigh myself.” In the bar graph, Darcy reflected that “maybe I can start tracking–measuring my steps, and see how many total steps that I’ve taken Sunday until Saturday.” And the line graph inspired Darcy to take it a step further by also recording “how much walk distance I walked.” The stacked bar chart told them “watch out for macronutrients ... like how much carb I’m eating. How much protein and fat I’m eating.” The visualizations also made Darcy want to tweak how they currently tracked their workouts: “this is telling to graph–to graph the time I’m exercising. When I’m tracking activities, I put the day that I’m exercising. I put the date and the type of the type of exercise I’m doing. But this graph is different, this graph shows you what time you are exercising ... I just put the date and then I put what I’m doing.”
4.4 A Broad Range of Abilities
Participants successfully completed more than half of the HDV reading activities (56.9%). Individual performances within each phase and across all three phases varied widely from person to person. While performance was generally better on the health data visualization identification and connection activities compared to the synthesizing skills used when reading beyond the data, some participants saw a smaller decrease in performance than others. Furthermore, the participants’ performance as they progressed through data visualization tasks indicated different abilities, strengths, and potential accessibility requirements across our sample (see Table
5). The variation to performance, like other IDD populations, indicate that the skills of people with Down Syndrome are highly heterogeneous when reading and making sense of HDVs.
Participants who were employed had a higher mean success rate. Skyler had the highest mean success rate across all three stages at 81.8%. Similarly, Emery, who uses a computer in their office job, was the most consistent, ranging between 58.3% and 71.8% success rates and the lowest overall performance decrease (11%). Emery was also the only participant to improve during the second connection phase from the first stage’s identification tasks. Other factors outside of those collected or observed in this study may have also affected performance.
While education in how to read a graph is useful, it was not the sole indicator of success and showed mixed results. As current high schoolers, Jordan and Cameron received the most recent instruction about data visualizations. Another student in secondary school, Sloane, did not mention any recent graph education. While Cameron had the 3rd highest mean success rate across all three stages (70.6%), Sloane was in the middle at 5th (54.1%), and Jordan scored the lowest at 38.3%. Darcy, a current university student who is passionate about their health, had the most consistent high success rate ranging from 68.8% to 79.3% and had the second highest overall mean score of 74.6%.
Although education may be a factor, resilience in response uncertainty and adaptability was a better predictor of a higher mean success rate. Participants in the lower half were among those more likely to abandon tasks or respond that they did not know how to answer or proceed. This more consistent trend across the participant pool indicated a barrier when they were uncertain what to do next, what elements meant, or how to connect elements together. Interestingly, this included Harper, the only participant who lives on their own. While Harper regularly demonstrates resilience and adaptability in their everyday life and throughout the study, this did not extend to tasks when they were unsure how to solve a task with more abstract data in a health context.
Although people with Down Syndrome have strong visual and spatial reasoning skills, as evidenced by our participant’s high performance identifying and comparing areas of HDV elements (\(\sim\)80% success rate), individuals like this study’s sample may similarly struggle making inferences and constructing understanding with how the information was connected. This appeared to occur when they were uncertain how to proceed or design elements forced them to make cognitive leaps during an HDV activity.