This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

\onlineid

0 \vgtccategoryResearch \vgtcpapertypeplease specify \authorfooter Samuel Reinders, Matthew Butler, Ingrid Zukerman, Lizhen Qu, and Kim Marriott are with Monash University. E-mail: [email protected] Bongshin Lee is with Yonsei University, E-mail: [email protected]

When Refreshable Tactile Displays Meet Conversational Agents: Investigating Accessible Data Presentation and Analysis
with Touch and Speech

Samuel Reinders    Matthew Butler    Ingrid Zukerman    Bongshin Lee    Lizhen Qu    and Kim Marriott
Abstract

Despite the recent surge of research efforts to make data visualizations accessible to people who are blind or have low vision (BLV), how to support BLV people’s data analysis remains an important and challenging question. As refreshable tactile displays (RTDs) become cheaper and conversational agents continue to improve, their combination provides a promising approach to support BLV people’s interactive data exploration and analysis. To understand how BLV people would use and react to a system combining an RTD with a conversational agent, we conducted a Wizard-of-Oz study with 11 BLV participants, where they interacted with line charts, bar charts, and isarithmic maps. Our analysis of participants’ interactions led to the identification of nine distinct patterns. We also learned that the choice of modalities depended on the type of task and prior experience with tactile graphics, and that participants strongly preferred the combination of RTD and speech to a single modality. In addition, participants with more tactile experience described how tactile images facilitated a deeper engagement with the data and supported independent interpretation. Our findings will inform the design of interfaces for such interactive mixed-modality systems.

keywords:
Accessible data visualization, refreshable tactile displays, conversational agents, interactive data exploration, Wizard of Oz study, people who are blind or have low vision.
\teaserThree images are included in this figure. The first, (a) shows an image of a bar chart that has been authored for display on a RTD. The second, (b) shows a photo of a RTD with that same bar chart displayed on it. The third, (c) shows a users hand touching the bar chart displayed on the RTD, with speech bubbles highlighting an interaction. The user asks the system -- what is the trend of this graph -- while the system responds -- the water storage begins at 83 percent. The three images are used to showcase how a system that combines refreshable tactile displays and conversational agents might be used to assist blind users with data understanding.

Using a Wizard-of-Oz method, we explored how refreshable tactile displays (RTDs) can be combined with conversational agents to assist people who are blind or have low vision (BLV) in undertaking data analysis activities. (a) Bar chart showing water storage over a 27-year period; (b) That same bar chart rendered on the RTD; and (c) BLV users could touch the RTD, perform touch gestures, and ask the conversational agent questions to aid their data understanding.

Introduction

A recent focus of data visualization research has been on how to best support people who are blind or have low vision (BLV) in their efforts to access visualizations. The research community has paid special attention to automated descriptions of visualizations and conversational interfaces [61, 2, 32], while the sonification community has focused on using non-speech audio to provide accessible data representations for BLV people [64, 65].

Raised line drawings, called tactile graphics, are widely used in the education of BLV students and in orientation and mobility (O&M) training. However, traditional printed tactile graphics are expensive to produce and unsuited for interactive usage such as data visualization and analysis. Fortunately, this is about to change, as low-cost refreshable tactile displays (RTDs) come onto the market [49]. RTDs are the equivalent of a computer display for BLV people. RTDs eliminate the recurring costs associated with traditional tactile graphics: the presentation of a tactile graphic is essentially free after the initial purchase of the device. Furthermore, RTDs support the interactive display of tactile graphics, as an image can be refreshed in only a few seconds.

However, the use of an RTD in isolation has disadvantages. RTDs currently have low resolution (making it difficult to provide braille labels), tactile graphics without an accompanying textual or verbal description are difficult to comprehend [44], and there are large disparities in the level of tactile literacy in the BLV community. This suggests that a conversational interface may complement an RTD, and that the combination of an RTD and conversational agent might overcome many of the current barriers BLV people face when analyzing data. This combination, however, has not been previously considered.

Our principal contribution is an initial understanding of how BLV people would use and react to such a multimodal system, from a Wizard-of-Oz study conducted with 11 BLV participants. In our study, participants were introduced to an RTD and were told that they could use direct touch input and gestures to control the RTD and engage in speech interactions with a virtual assistant (i.e., a text-to-speech interface managed by the wizard). They were given three data analysis scenarios, one for training, and completed several tasks in each scenario, which they could address using the RTD and/or the speech interface as they wished. As these scenarios focused on basic data understanding and analysis, we chose to present time series data using line charts and bar charts, and spatial data using isarithmic maps. From the analysis of participant behaviors, we identified preferred modalities and patterns of interaction and how these were influenced by participants’ tactile experience and task type (e.g., identify trend, find values).

The main findings of our study are as follows:

  • Almost all of our participants indicated that using a combination of an RTD and conversational agent to analyze data had benefits over single formats (i.e., RTD only or agent only) or traditional tactile graphics alone (Section 4.1).

  • Depending on the task type, participants gravitated towards different patterns of interactions. For example, touch was crucial during initial exploration, while touch gestures and speech were more frequently used when identifying values or extrema (Section 3.2).

  • Participants’ level of tactile experience influenced their interaction patterns (Section 3.2.2), and more experienced tactile graphic users spoke of how the RTD enabled a deeper engagement with the data and independent interpretation (Section 3.4).

Our findings highlight the desire by BLV people to engage in independent and meaningful data exploration, and indicate that this is supported by the combination of an RTD and conversational agent. Our work provides the basis for the future development of such systems.

1 Related Work

1.1 Accessible Data Visualization

The growing availability of data has allowed decision-making to become increasingly data-driven in virtually all professions. Employees are now expected to use tools like Excel or Tableau for data-driven decision making, and to employ a mix of analytics and visualization. The use of data visualization is, however, problematic for BLV people, and we are witnessing a surge of research interest in investigating accessible alternatives (e.g., [36, 33, 38, 37, 17, 60, 74]).

One approach is to provide a written or spoken description of a graphic. Researchers have investigated what makes a good description [29], how to automatically generate a description from a graphic or answer questions about a graphic (see [18, 26] for recent reviews), or how to generate descriptions directly from data [70, 50, 51]. Some researchers have explored how to incorporate these capabilities into systems designed to support data analysis by BLV people [61, 2, 32].

Another approach is to use non-speech audio, known as sonification, to convey information. For instance, the SAS Graphics Accelerator uses sonification and tables to provide access to visualizations [58]. Sonification can be used as an alternative format to data visualizations, or alongside data visualizations [35]. Recently, there has been interest in combining speech with sonification [24, 64, 65].

The third approach is to provide raised line drawings known as tactile graphics, which are encountered by many BLV students at school. Transcription guidelines recommend their use for accessible provision of maps, diagrams, and graphs, rather than textual or verbal descriptions, in particular if the reader needs the information to participate in discussions or complete a task [44]. Most tactile graphics are produced using swell paper or embossing [56]. However, printed tactile graphics are not well-suited for interactive data exploration, because of the cost and speed of production. Therefore, there has been considerable research into displays that combine auditory and tactile feedback as alternatives to tactile graphics, which include touch screens with audio labels and with/without vibration feedback (e.g., [20, 40]). However, previous research suggests that exploration is faster and more accurate with tactile graphics than with touchscreen-based approaches [8].

1.2 Refreshable Tactile Displays

A promising alternative is large refreshable tactile displays (RTDs) [66, 69]. Most consist of a grid of pins controlled by electro-mechanical actuators [69]. Their great advantage over traditional tactile graphics is that it takes only a few seconds to display a new graphic. In other words, an RTD is the tactile equivalent of a computer monitor, and like a monitor, it supports interactive data exploration.

Until recently, RTDs have been prohibitively expensive. However, the DAISY Consortium’s Transforming Braille Project has spurred the development of cheaper devices, which are now entering the market. These include the DotPad [13] and Monarch [3], which provide 2,400 (60×\times40) and 3,840 (96×\times40) pins. Other devices on the market include the Graphiti [46] (2,400 (60×\times40) pins) and Metec [39] (6,240 (104×60104\times 60) pins), which offer several pin heights and touch-sensitive pins.

Research into the use of RTDs has focused on the display of static images in the fields of art [21] and book diagrams and illustrations [34, 43, 48]. Other work has looked at the display of dynamic images, such as sports matches [45] and animations [22]. RTDs have also been used to show maps for building a cognitive model pre-travel [6, 28, 41, 59, 71], or as mobility aids updated in real-time to show position/obstacles as a BLV person moves through an environment [5, 72, 73].

There has been almost no research into the use of RTDs for interactive data visualization. One paper describes the co-design of tools for the interactive exploration of set diagrams and parallel vectors on an RTD [14], and a poster at VIS 2023 examined stakeholder perspectives of how RTDs could be used to improve access to graphics, including data graphics [23]. There has been more research into systems that use RTDs with audio labels to display maps automatically generated from web-based map systems, which support search, zooming, panning, and route finding (e.g., [59, 73, 71]).

1.3 Multimodal Display of Graphical Information

A limitation of RTDs is their low resolution when compared with traditional tactile graphics [47, 22]. This makes the provision of braille labels problematic, and is one of the reasons why audio labels are provided on most RTD map applications (e.g., [59, 73, 71]). Indeed, in their review, Butler et al. [8] found that audio labels on tactile graphics are generally preferred to braille labels and lead to faster performance. Furthermore, it is recommended that tactile graphics be accompanied by a spoken or written description [44].

Advances in question-answering interfaces and conversational agents suggest that we can go one step further and combine an RTD with a conversational agent. We know of no prior research into such a combination. However, there has been some research into combining conversational agents with tactile maps [9], tactile graphics [19], art pieces [27, 10], and with 3D printed models [62, 54, 53]. Shi et al. [62] determined that in addition to audio labels, 3D printed models should allow BLV users to ask questions using conversational language. Reinders et al. [54] found that BLV participants desire multimodal interactions that combine spoken dialogue, touch gestures, and haptic vibratory feedback; this finding was followed by design recommendations that heavily focused on conversational interactions [53].

1.4 Wizard-of-Oz Method

A Wizard-of-Oz (WOz) study [12] is a research method where an end user interacts with an interface that, to some degree, is being operated by a ‘wizard’ who fulfills an interaction functionality that is yet to be fully implemented [31]. WOz studies often involve eliciting interactions from end users, with the wizard simulating system behaviors that complete their interactions [68, 11, 63]. WOz enables researchers to observe the use of speculative interfaces or artefacts, the outcomes of which can influence future designs.

WOz studies have been used in various domains and contexts. Salber and Coutaz [57] demonstrated how WOz can be extended to the analysis of multimodal interfaces, and formulated a set of requirements for a generic multimodal WOz platform. Kahn et al. [30] employed a WOz technique to control some of a humanoid robot’s speech and actions to investigate patterns in human-robot interactions. In information visualization research, Walny et al. [67] exploited the WOz method to investigate pen and touch interactions for chart creation and manipulation on interactive whiteboards. Recently, researchers applied WOz to investigate BLV users preferred interaction preferences with multimodal systems combining 3D printed models and conversational agents [63, 54]. Kim et al. [32] conducted a WOz study to understand if and how Question Answering systems can help BLV users’ visualization comprehension.

Our WOz study explores a system that combines a conversational agent with an RTD to support BLV people’s interactive data analysis. Our objective is to propose initial design guidelines and to meaningfully engage BLV people in the design of such a system.

(a) shows a line graph depicting the price of a stock increasing over time. It has axes and a braille title. (b) is a line graph showing water storage data. It has axes, a braille title and a legend label -- water. (c) is a line graph showing water rainfall data. It has axes, a braille title and a legend label -- rain.
(a) Stock line graph (b) Water storage line graph (c) Water rainfall line graph
(d) is a multi line graph that shows water storage and water rainfall overlapping. It has axes, a braille title and two legend labels -- rain and water. (e) showcases water storage data transformed into a bar graph. It has axes and a braille title. (f) is an isarithmic map of Australia graphing fire weather data. It has a braille title and four legend labels.
(d) Water storage and rainfall multi line graph (e) Water storage bar graph (f) Fire isarithmic map
Figure 1: Six of the data visualizations used in the user study. For Water, participants could choose between rainfall or water storage data, or both together. The individual Water line graphs could also be presented using bar graphs (e.g., (e)) that represent the same information.

2 User Study – Methodology

We conducted a user study (1) to evaluate the perceived usefulness of the combination of an RTD and a conversational agent to support data analysis by BLV people and (2) to identify the ways in which BLV people wish to interact with such a combination for data analysis. A particular focus was on understanding how they wished to use the different modalities of touch, touch gesture, and speech.

To achieve these goals, we used a WOz method to simulate system behaviors and elicited user interaction needs. This allowed us to capture how the BLV participants chose to interact with an RTD and conversational agent when conducting data analysis free of technological constraints. We also conducted a questionnaire and semi-structured interview to further explore participant responses to the system.

2.1 Participants

We recruited 11 BLV participants (Table 1) from our lab-managed participant contact pool and through mailing lists of BLV support and advocate groups. Participants were aged between 21 and 77 (μ\mu = 44, σ\sigma = 17.2). Eight participants identified as totally blind, while three were legally blind. Onset of vision loss was congenital (n=7n=7) or occurred during childhood (n=1n=1), adolescence (n=2n=2), or adulthood (n=1n=1). Participants varied in their experience with tactile graphics: five reported substantial exposure to and confidence using tactile graphics, and three reported some use but lacked confidence, while others reported either limited (n=1n=1) or no exposure (n=2n=2).

Table 1: Participant demographic information, detailing age, level of blindness (TB = totally blind, LB = legally blind), onset of vision loss, and self-rated competency using Tactile Graphics (TG).
Part. Age Blindness Onset TG Experience
P1 21 LB Congenital Substantial
P2 25 TB Congenital Substantial
P3 42 TB 5 y.o. Substantial
P4 54 TB Congenital None
P5 33 LB 15 y.o. Limited
P6 53 TB Congenital Some
P7 45 LB 15 y.o. Some
P8 40 TB Congenital Some
P9 77 TB 45 y.o. None
P10 65 TB Congenital Substantial
P11 29 TB Congenital Substantial

All participants regularly used conversational interfaces, including Google Assistant (n=10n=10), Siri (n=9n=9), and Alexa (n=5n=5). These interfaces were commonly accessed on smartphones (n=11n=11) and smart speakers (n=10n=10), followed by smart watches (n=7n=7), computers (n=7n=7), tablets (n=5n=5), and television (n=3n=3). Additionally, several participants (n=5n=5) mentioned some use of ChatGPT.

Participants self-reported their technology comfort level using the Technology Adoption Cycle scale [55]. Most (n=10n=10) were Early Majority and one was Late Majority. All other categories, Innovator, Early Adopter, and Laggard, were absent.

Participants were also asked to self-report their level of comfort with basic and complex data manipulation/analysis tasks on a five-point Likert scale. Most participants were comfortable with basic tasks (e.g., creating a budget): Strongly Agree (n=5n=5), Agree (n=4n=4), and Neutral (n=2n=2). Comfort with complex tasks (e.g., performing a statistical analysis) was more mixed: Strongly Agree (n=1n=1), Agree (n=4n=4), Neutral (n=3n=3), Disagree (n=1n=1), and Strongly Disagree (n=2n=2).

2.2 Materials

For the study we used Graphiti [46] (Figure When Refreshable Tactile Displays Meet Conversational Agents: Investigating Accessible Data Presentation and Analysis with Touch and Speech(b)). The Graphiti’s display has 2,400 (60×\times40) pins, each of which can be actuated to one of four heights. The Graphiti includes keys to navigate between graphics, and the display can render Braille text. When connected to a computer, a PC application can be used to create or adjust graphics, with changes appearing in real-time on the RTD.

As our work is the first utilizing an RTD for data presentation, we decided to focus on basic data understanding. We selected univariate time series and spatial data, and visualization types we believed our participants would be most familiar with. We selected datasets from finance and climate–Stock, Water  and Fire–that we believed would be both timely and of interest to participants [32], and created several visualizations (line charts, bar charts, and an isarithmic map) (Figure 1).

  • Stock gives the price of a stock over four and a half years (Figure 1(a)). It was presented using a line chart, with years on the horizontal axis and price on the vertical axis. This was shown first to familiarize participants with the device and its capabilities.

  • Water gives the rainfall and water storage levels in the state of Victoria, Australia between the years 1995 and 2022. The water storage and rainfall information were presented as line charts, either separately (Figure 1(b)-(c)) or together as a multi-line chart with dual axes (Figure 1(d)), depending on the participants’ preference.111We are aware that dual axes time series may be visually misleading, though for a counter-argument see [15]. However, the limited resolution of the RTD effectively precludes the use of line chartss stacked above one another, which is the most commonly recommended alternative. We therefore decided to provide our participants with the choice.

    Participants could also request to change the presentation format from line chart to bar chart (Figure 1(e)). Line charts are better suited to showing trends, while bar charts are better suited to discrete comparisons. We therefore decided to allow participants to change the visual encoding.

  • Fire gives changes in the number of days with dangerous fire weather across Australia between the years 1951 and 2022, using an isarithmic map (Figure 1(f)). Data points are charted using four pin heights on a map of the country: Height 1 (the shallowest) represents a yearly decrease in fire weather, Height 2 represents no change, Height 3 a yearly increase of between 1 and 10 days, and Height 4 (the tallest) an increase of between 11 and 20 days.

All materials were developed with feedback from an experienced tactile transcriber. All visualizations included a braille title and legend. Due to the limited resolution of the Graphiti (in common with other RTDs), braille axis labels and markers were not provided.

2.3 Study Setup and Procedure

User study sessions took place at a location convenient to each participant, either at their residence or at the researchers’ university campus. One or two researchers were present during the experiment.

2.3.1 The Role of the Wizard

A member of the research team took the role of the wizard during user study sessions. The wizard acted on behalf of the RTD device and conversational agent in response to participants’ interactions. The wizard used the Graphiti PC application to adjust graphics in real-time (e.g., zooming or changing which visualization was displayed). To provide feedback to participants quickly and consistently, the wizard answered questions posed to the conversational agent by selecting responses from a script created for each visualization. In line with Kim et al.’s work [32], each script covered the visualization title, overview, axes values, data values for each data point, trend, extrema (minimum and maximum), and background information regarding the visualizations (e.g., “what is a bar chart,” “what constitutes a dangerous fire weather day”). Responses were generated on the fly using text-to-speech, and played back using a connected speaker. If a particular request was outside of the scope of a script and could not be satisfied (e.g., converting a line chart into a scatter plot), the wizard would inform the participant that the functionality in question had not been implemented and would respond: “I currently don’t understand how to answer that, as I am a prototype, but I will note your request down for the future.”

In line with typical WOz studies, we informed our participants of the true nature of the WOz deception at the conclusion of the study session. However, two participants (P2 and P3, having had previous experience with WOz studies) asked during the study if the interface was “real,” and they were informed of the true nature of the interface before continuing with the study.

2.3.2 Survey and Overview

Sessions began with a brief questionnaire that included questions about participant demographic information, accessible format use, technology use, and data proficiency. Then, participants were informed that they would be interacting with a prototype device that combines an RTD and a conversational agent to help them explore data and complete basic data analysis tasks. Participants were asked to verbalize any actions they wanted to perform using the ‘think-aloud’ protocol, as the focus of the session was on observing their chosen ways of interacting with the device. This allowed the wizard to act on behalf of the RTD and conversational agent when necessary. Think-aloud has been used widely in elicitation studies [68], including those that utilize WOz experiments in the accessibility space [63, 54, 32].

2.3.3 Familiarization with the System

The Stock visualization was used to familiarize participants with the Graphiti device, the think-aloud protocol, and the ways they might interact with the device. The researcher first provided an overview of the device and its technical capabilities. This included outlining that it could generate graphics by raising small pins arranged in 40 rows of 60 pins, and that it supported natural language understanding and generation. Participants were provided with example interactions they might choose to use, e.g., touch gestures, asking questions with speech using the wake phrase “Hey Graphy,” and using the RTD’s buttons, either in isolation or in combination. Participants were then asked to explore the Stock visualization.

Once they finished their exploration of the Stock visualization, participants were directly elicited to choose modalities and interaction techniques to perform a series of basic tasks: (1) getting the title of a graphic, (2) getting an in-depth description of a graphic, and (3) selecting a data point and extracting its value (detail on demand). The wizard acted on behalf of the RTD to fulfill the participants’ requests.

2.3.4 Interacting with Water and Fire Visualizations

Participants were first introduced to the Water visualization, followed by the Fire visualization.

Water. The researcher explained that participants could choose to display water catchment storage, rainfall, or both types of data together: the appropriate single or multi-line chart was displayed on the RTD based on the participants’ choice.

Participants were then asked to explore Water. Upon the first touch, the agent provided a spoken overview of the Water graph. Participants were also invited to interact with the visualization to perform the following operations: (1) zooming, (2) panning, (3) adding/removing data sets, (4) filtering, (5) transforming to another representation (line chart to bar chart), and (6) undo. Again, elicited interactions were fulfilled by the wizard.

Participants were then asked to complete several data analysis tasks using either the storage line chart, rainfall line chart, their respective bar charts, or the combined storage and rainfall multi-line chart.

Fire. Participants were asked to explore the Fire isarithmic map visualization, with the agent providing a spoken overview when first touched. Participants were then asked to perform similar analysis tasks to those completed with Water.

Analysis tasks. For both Water and Fire, participants were asked to complete three types of tasks, representative of identification, comparison, and summarization [42]. Our choice of tasks was also influenced by other systems aimed at facilitating data understanding by BLV users [61, 26].

  • Identifying specific values: for the Water visualization, this task involved reporting the most recent storage/rainfall, and that in 2005, and for Fire, this task required identifying the number of fire-weather days in Brisbane, Australia.

  • Identifying extreme values: for the Water visualization, this task involved reporting on the lowest and highest storage/rainfall, and for Fire, participants had to find an area where the number of fire-weather days had decreased.

  • Describing the trend: this task required participants to report on the general trend of the phenomenon of interest in both visualizations.

Participants were asked to complete these tasks using any interactions or transformation operations they found suitable. Afterwards, they were given the opportunity to further explore Water and Fire, and asked to articulate any interesting facts that they had identified.

2.3.5 Interview

The session ended with a semi-structured interview. The researchers asked questions to better understand why participants used particular interaction modalities, their comfort level during interactions, particular use cases, whether the combination of an RTD and conversational agent would make them more confident undertaking data tasks, and if the combination was preferred to a conversational agent, RTD, or tactile graphic in isolation. Sessions lasted between 1.5 to 2 hours, and participants were compensated with a $100 AUD gift card.

2.4 Data Collection and Analysis

User study sessions were video and audio recorded. All captured footage was transcribed.

Our aim was to capture and categorize the way in which our BLV participants chose to interact with the system. To achieve this goal, we classified each interaction based on the combination of modalities used (touch, conversation, or touch gestures), as well as their order and intent. One researcher derived an initial set of patterns by classifying the interactions of the first user study session (P1), which were discussed and confirmed with a second researcher. This set was supplemented with additional patterns identified in subsequent user study sessions (P2 - P11). Whenever new patterns were added, all previous sessions were reassessed. On completion, the second researcher validated a subset of the classified interactions, with a subsequent meeting held to reconcile conflicts. The final set of interaction patterns is presented in Section 3.

A thematic analysis was performed similarly on (1) comments made by participants during a session and (2) participant responses to the semi-structured interview. An initial set of codes was devised based on the researchers’ observation of the first user study session (P1). Two members of the research team independently coded this session and met to compare codes, which were then refined and extended. All subsequent user study sessions were then coded by one researcher, with the second researcher validating a subset of the coded data. An additional meeting was held to discuss and reconcile any coding conflicts. Whenever new codes were added, combined or modified, all previous sessions were reassessed. The complete set of coded data was examined and analyzed to generate a set of themes. Three researchers met to review the themes, which were then refined and consolidated. The final set of themes is described in Section 3.4.

3 Results

We present participant command interaction preferences, modality interaction patterns, participant conversational queries, and a thematic analysis of participant responses.

3.1 Command Interaction Preferences

Multiple participants controlled the system (referred to from now on as ‘Graphy’) using commands based on prior experience with devices, including smartphones, smart speakers, and computers.

To access the visualization title, six participants preferred to touch read a braille label, e.g., “I would touch the top and feel for braille” (P6). Graphy provided an overview of each visualization upon first touch; to repeat the overview, seven participants asked the device, e.g., “Hey Graphy, can you describe the chart?” (P3).

When extracting detail on demand, participants either performed touch gestures (nn = 55), e.g., single and double tapping to extract descriptions of values at the location at which the gesture was performed, or combined touch gestures with conversation (n=6n=6), e.g., tap and hold on a point and ask a question, expecting the device to use their location in conjunction with the question.

Table 2: Modalities that participants used when performing basic functions and data operations (T = Touch, C = Conversation, G = Touch Gesture, K = Keyboard, G+C = Touch Gesture and Conversation); modalities used in more than 20% of each task/operation are highlighted.
Operation Modalities
T C G K G+C
Visualization Title 6 3 2 0 0
Visualization Overview 3 7 1 0 0
Detail on Demand 0 0 5 0 6
Zooming 0 0 6 1 4
Panning 0 4 6 0 1
Adding/Removing Data 0 8 0 3 0
Filtering 0 8 1 2 0
Transforming 0 9 1 1 0
Undo 0 9 1 1 0

For zooming, six participants used the pinch-to-zoom gestures used on their smartphones, e.g., “I am so used to the iOS gestures, can I do that?” (P8), while four combined gestures with voice, e.g., “Hey Graphy, zoom on the area I am holding” (P7). For panning, six participants used swiping gestures from smartphones, e.g., swiping to the left or right with one or more fingers.

For operations that do not cleanly map to smartphone gestures, participants gravitated towards asking the device, or in some cases, using the RTD’s keyboard. Most participants filtered, transformed, added/removed data, and reversed operations through conversation, e.g., “Hey Graphy, only show areas of the isarithmic map where fire risk has decreased” (P3). Several participants chose to use the RTD’s braille keyboard, some of whom indicated that it would be better to be able to connect a QWERTY keyboard, e.g., “I want to type \ldots like when I interact with ChatGPT” (P6).

3.2 Modality Interaction Patterns

All participants combined touch, touch gestures, and natural language queries to interpret and analyze the data presented in the visualizations (Figure 2, Table 2).

An image showing two participants interacting with visualizations rendered on the refreshable tactile display. In the first image, a participant has their hands tracing a line chart that has been rendered on the display. In the second image, a participant is touch reading braille labels that belong to a legend associated with an isarithmic map of Australia that is rendered on the display.
Figure 2: Participants interacting with Water and Fire visualizations.
Table 3: Modality interaction patterns specified; the modalities involved in each pattern (T = Touch, G = Touch Gestures, C = Conversation).
Modality Interaction Pattern Modalities
T C G
Conversation Only *
Conversation and Undirected Touch * *
Conversation followed By Directed Touch * *
Directed Touch Only *
Directed Touch followed By Conversation * *
Directed Touch followed By Touch Gesture * *
Interleaved Directed Touch and Conversation * *
Interleaved Directed Touch and Touch Gesture * *
Interleaved Directed Touch, Conversation, and Touch Gesture * * *
Table 4: Modality interaction patterns for four task types (exploration, identifying specific values, identifying extreme values, and describing a trend); Touch is directed unless otherwise specified; number of times each pattern was used, percentage for each task type, and number of unique participants that performed each pattern during each task type; finally, the top two patterns employed during each task type are highlighted.
Task Types
Interaction Modality Pattern Exploration Specific Values Extreme Values Trend
# times % # participants # times % # participants # times % # participants # times % # participants
Conversation Only 0 0% 0 2 6% 2 3 9% 2 2 10% 1
Conversation and Undirected Touch 0 0% 0 2 6% 2 3 9% 3 1 5% 1
Conversation followed By Touch 0 0% 0 1 3% 1 0 0% 0 4 19% 3
Touch Only 16 49% 8 0 0% 0 2 6% 2 9 43% 6
Touch followed By Conversation 0 0% 0 8 24% 6 7 21% 6 1 5% 1
Touch followed By Gesture 0 0% 0 6 18% 3 2 6% 1 0 0% 0
Interleaved Touch and Conversation 10 30% 5 10 30% 7 10 30% 7 2 9% 2
Interleaved Touch and Gesture 7 21% 3 3 9% 2 5 15% 4 2 10% 1
Interleaved Touch, Conv., and Gesture 0 0% 0 1 3% 1 1 3% 1 0 0% 0
Total 33 100% 33 100% 33 100% 21 100%

We identified nine distinct modality interaction patterns that were used during exploration and analysis tasks based on the modalities – Touch, Touch Gesture, or Conversation – and the order in which these modalities were used (Table 3). We distinguished between directed touch, where the participants purposefully explored a graphic, and undirected touch, which was more haphazard, and appeared to lack a purpose. The choice of interaction pattern strongly depended on the type of task and prior tactile experience. The presentation format (line chart, bar chart, or isarithmic map) did not appear to influence the interaction pattern and choice of modality.

3.2.1 The Impact of Task

Table 4 shows the patterns that participants used for each task.

Initial exploration. The Water line charts were explored on average for 3:16 min, the Water bar charts for 1:19 min, and the Fire isarithmic maps for 3:07 min. The exploration of the Stock line chart was excluded from the results, as it was used for training.

All participants utilized touch in the initial exploration of the Water and Fire visualizations. About half (49%) of the exploration tasks were completed using Directed Touch Only, i.e., initial information gathering of the visualizations was performed purely through touch. Two other patterns, Interleaved Directed Touch and Conversation (30%) and Interleaved Directed Touch and Gesture (21%), were also used. While most time was still spent exploring via touch, five participants asked the conversational agent questions while exploring. For example, they asked for a description of the graphic or the difference between a line chart and a bar chart, and used speech to obtain descriptions of data values or axis labels before returning to touch.

Identifying specific or extreme values. Even though the answer could be easily obtained by directly asking the conversational agent for these values, the Conversation Only and Conversation and Undirected Touch patterns were infrequently used. Instead, 87% of the specific-value tasks and 81% of the extreme-values tasks were completed using touch with speech or gestures such as double taps: Interleaved Directed Touch and Conversation, Directed Touch followed by Conversation, and to a lesser extent Directed Touch followed by Gesture and Interleaved Directed Touch and Gesture.

When asked to identify specific values, e.g., the most recent rainfall in Water or the number of fire-weather days in Brisbane in Fire, participants generally built an initial understanding through touch, e.g., counting bars or x-axis markers to identify a data point and tracing it to the y-axis, before confirming values using either speech or a gesture (Directed Touch followed by Conversation [24%] or by Gesture [18%]). Across the three specific-values tasks, seven participants also chose to verify this output by undertaking further Directed Touch, employing the Interleaved Directed Touch and Conversation (30%) pattern, e.g., comparing the found value to surrounding data points, or re-tracing to corresponding y-axis points.

These patterns were used in similar ways when identifying extremum values, e.g., the highest water storage in Water or areas of the Fire isarithmic map that showed decreases in fire-weather days. Participants would often start with touch, by hovering and rubbing their hands in circular motions to identify extremum charted values, before asking the conversational agent for exact values. In many cases, participants would then trace between close extrema to compare them, undertaking subsequent spoken or gesture-based interactions to extract values.

Describing a trend. All participants were able to describe the general trend of the Water and Fire visualizations. Just under half (43%) of the trend tasks were completed using Directed Touch Only to build up an understanding of the visualizations. This included tracing the Water line chart or bar chart, identifying key extrema, and comparing values in a systematic way. For Fire, this involved tracing the isarithmic map’s perimeter and rubbing hands in circular motions over the isarithmic map to identify areas with change. One participant (P5) described the Fire trend verbally, without any pattern or explicit interaction (touch or otherwise) altogether, instead relying on the knowledge they had built during earlier tasks.

The multimodal pattern Conversation followed by Directed Touch was also used, but to a lesser degree (19%). Three participants asked the conversational agent to describe a trend, followed by touching the visualizations to compare and seemingly validate the agent’s answer.

3.2.2 The Impact of Tactile Experience

The participants’ self-reported tactile experience appeared to influence which interaction patterns they used when completing tasks (Table 5). One participant (P5) reported limited exposure to tactile graphics, and two (P4 and P9) reported no exposure (Table 1). P9 spoke of how a lack of experience made them hesitant to rely on touching the visualizations, “… it takes me a while, I suppose because I am not really used to tactile.” These participants appeared to consume information through touch in a way that had less definition, and at times used touch in a way that was undirected (goalless) or unsuccessful. P9 described difficulty visualizing what they were touching – “I can’t grasp what this is telling me very well.” At the same time, P4 confused the isarithmic map for a bar chart, and P5 experienced difficulty locating a prominent island featured on the isarithmic map. There were also cases where these participants touched the RTD mid-refresh, which caused confusion.

Participants with low tactile experience (n=3n=3) were the only group that relied on the Conversation Only (19%) and Conversation and Undirected Touch (16%) patterns in a significant way (Table 5). In these interactions, the low tactile experience participants were satisfied just by asking the conversational agent for answers.

Participants with medium (n=3n=3) and high (n=5n=5) tactile experience were more comfortable completing tasks using Directed Touch Only, choosing this modality to complete roughly a quarter of the tasks, and using touch with conversation and/or gestures to complete the remaining tasks. The only noticeable difference between these two groups was that participants with medium tactile experience preferred to use the conversational agent over gestures, whereas the more experienced participants tended to switch between them.

Table 5: QA strategy use classified by self-reported level of tactile experience (Low [Limited/None] = P4, P5, P9; Medium [Some] = P6, P7, P8; High [Substantial] = P1, P2, P3, P10, P11); the top two patterns employed by each group, based on tactile experience, are highlighted.
Tactile Experience
Interaction Modality Pattern Low (n=3n=3) Med. (n=3n=3) High (n=5n=5)
# times % # times % # times %
Conversation Only 6 19% 0 0% 1 2%
Conversation and Undirected Touch 5 16% 1 3% 0 0%
Conversation followed By Touch 2 6% 2 6% 1 2%
Touch Only 5 16% 8 24% 14 26%
Touch followed By Conversation 2 6% 7 21% 8 15%
Touch followed By Gesture 2 6% 0 0% 6 11%
Interleaved Touch and Conversation 6 19% 13 39% 12 22%
Interleaved Touch and Gesture 4 13% 1 3% 12 22%
Interleaved Touch, Conv., and Gesture 0 0% 1 3% 1 2%
Total 32 100% 33 100% 55 100%

3.3 Conversational Queries

Participants asked Graphy’s conversational agent a total of 202 queries (μ=18.5,σ=15.9\mu=18.5,\sigma=15.9), across four categories:

  • Data Extraction (102, 50%) queries related to the retrieval of data values, e.g., “what is the value [that I am touching]?” or “what was the rainfall in 2006?”

  • Operations (69, 34%) queries to manipulate the data or visualizations, e.g., “can you zoom this area [of the graph]?” or “can you transform this data into a bar graph?”

  • Overview (20, 10%) queries to summarize the data, e.g., “describe this chart” or “what is the trend showing on this chart?”

  • Understanding Visualization (11, 5%) queries that provide context relating to the type, layout or encoding of a visualization, e.g., “how many segments are in this graph?” or “what is a bar graph?”

3.4 Semi-Structured Interview and Thematic Analysis

We asked participants to indicate their level of agreement with statements about whether the combination of the RTD and the conversational agent would make them more comfortable completing basic data analysis tasks and complex tasks, and also whether this combination offers significant benefits over each format alone—conversational agent, RTD, or tactile graphic (5-point Likert scale). We ran binomial tests to determine whether the difference between the number of participants who agreed or strongly agreed and those who were neutral or below is statistically significant. We performed Holm-Bonferroni corrections for multiple comparisons [25].

Most participants agreed or strongly agreed with the statement about basic data analysis tasks (k=8k=8, n=11n=11, p=0.029p=0.029), and more participants agreed or strongly agreed when asked about more complex tasks (k=10k=10, n=11n=11, p=0.003p=0.003). In addition, participants showed a strong preference for the combination over just a conversational agent (k=11k=11, n=11n=11, p<0.001p<0.001), RTD (k=9k=9, n=11n=11, p=0.012p=0.012), or tactile graphic alone (k=10k=10, n=11n=11, p=0.003p=0.003) (Figure 3).

A bar chart representing participants’ views on a 5-point Likert scale (from Strongly Disagree to Strongly Agree) regarding the benefits of the RTD plus the conversational agent compared to just a conversational agent alone, just the RTD alone, and a tactile graphic alone. The bar that presents a comparison with just a conversational agent shows that all 11 participants answered Strongly Agree. The bar that represents comparison with just the RTD shows 8 Strongly Agree, 1 Agree, and 2 Neutral. The bar that presents a comparison with just a tactile graphic shows 7 Strongly Agree, 3 Agree and 1 Neutral.
Figure 3: Participants’ views on a 5-point Likert scale (1: Strongly Disagree and 5: Strongly Agree) regarding the benefits of the RTD plus the conversational agent (CA) compared to the CA alone, the RTD alone, and tactile graphics (TG) alone; the x-axis represents number of participants.

A thematic analysis of participant responses during the WOz study and interview provided further insights, yielding the following themes: usefulness and use cases, multimodality, deeper experiences through tactile visualization, autonomy, user choice, and design feedback.

Usefulness and Use Cases. All 11 participants described ways they could integrate a device that combines an RTD and a conversational agent into their everyday lives, both work and personal.

P8 described how it would help “reduce the fear and anxiety of [working with] graphs and charts,” suggesting they would like to use Graphy to access the railway heat maps they encounter at work. P1, who works as a data analyst, noted that the device would “make their life a whole lot easier,” and suggested that it could be used to explore stock prices and perform technical analyses. P11 indicated that Graphy could display many of the charts and tables used in payroll duties.

Five participants shared stories of being under-equipped to participate in data literacy activities during schooling, and six suggested that access to a device combining an RTD and a conversational agent could have expanded tertiary opportunities, including statistics (P4, P1, P8), mathematics (P5), computing algorithms (P2), and psychology (P10).

Seven participants spoke of how Graphy could improve access to data for personal uses. This included personal finance and budgeting (P3, P7), monitoring and interpreting health data, such as blood sugar and blood pressure (P7), real-time sports data (P6), and weather forecasts (P9). P7, who became blind at 15, described the utility of Graphy well, “This is great. I would love a real world application of this. There is a lot of data I don’t have access to that I did say 10 years ago.”

Multimodality. Four participants described how Graphy’s multimodal nature was complementary. P8 spoke of how both the RTD and conversational agent were integral to their understanding of the visualizations, stating that “the RTD is tangible and the conversational agent provides the conceptual information. Touch is the strongest sense we have after sight to observe the physical world. The conversational agent then helps us to understand. The combination of both incorporates knowledge and experiment.” P3 described the multimodal combination as having “so many advantages because it is more of a 360 degree view of the information.”

P7 discussed how the conversational agent allowed them to “connect the dots between what [they were] touching,” while P6 preferred the combination that Graphy offered compared to printed tactile graphics, indicating that “the conversational agent allows me to get answers to my questions or have the confusion cleared up, while the printed page will remain silent and force me to guess [or move on].”

Deeper Experiences through Tactile Visualization. Several participants described how tactile graphics helped facilitate deeper experiences with data and allowed for independent interpretation.

The Fire isarithmic map, in particular, helped participants build a ‘visual image’ in their head. P8 felt that after the session, the map would “be stuck in my mind, I can still visualize [the state of] Queensland and Australia,” and continued, “my brain is [so] stimulated at the moment.” P8 later discussed how the data would have less meaning to them if it were presented in another format, like a table. P1 became visibly excited, while P7 expressed emotion touching the isarithmic map, describing “I love the shape of Australia \ldots that is something you really miss when you cannot see. Simple things like a map.”

Three participants (P1, P2 and P6), all of whom had higher tactile experience, indicated how the tactile graphics supported independent interpretation. They did not want to simply “be told the answer,” but instead preferred to build their own understanding of the data first through touch. P1 repeatedly acknowledged that they could just ask Graphy for results, but first wanted to explore the visualizations and work out their own answers. P6 described their process, “my first instinct is to use my hands to first feel what the graph is doing… I would like to find the answer myself, then use voice to verify it, you feel more accomplished if you find [it] yourself.” P2 even became visibly frustrated when Graphy gave too much information for their question, feeling that their independence of interpretation had not been respected – “I didn’t ask for an analysis… I just wanted to know what the data was… that would be the type of analysis I would [like to] do myself.”

Autonomy. Four participants mentioned how they valued that Graphy could provide independent access to data without the need for assistance. P8 described how “Graphy explains things better, otherwise [I would] require someone else who [can explain the concept],” while P3 felt that the device had allowed them to build an understanding equivalent to what a helper would, but on their own. P6 felt that Graphy made them more autonomous, “it feels more independent… I don’t need somebody to tell me the answer.” P5 spoke of how in high school they had a support worker who would create sine graphs out of modelling clay, and suggested that the combination of RTD and a conversational agent could serve a similar purpose.

User Choice. Participants mentioned that the ways they would choose to use Graphy depended on purpose, personal preference, environment, and prior experience with technology.

Three participants discussed how their use would depend on the purpose of the data analysis. P1 described that it would “depend on what I am doing… how deeply I want to explore the data,” stating that in an educational or work context they would engage in deeper interactions to ensure they completely understood what was graphed. P2 discussed varying interactions based on the effort required, “I am more likely to use the conversational agent for more complex tasks, [those where] the effort to benefit ratio is better,” while P8 stated that for basic tasks they would rely on touch, but for work tasks would use touch gestures and voice interaction.

Four participants mentioned that the way they interacted was a personal choice. P4 mentioned that “these are my needs, other people might think and interact differently,” while P5 stated that “the [methods] are all useful, but I prefer buttons and touch.” P3 connected their preference to level of ability, indicating that “because of paralysis, my hand can get sore, so designing it so I can use voice interaction is really important,” while P8 favored voice interaction as “it is what I am most comfortable with.”

Unsurprisingly, all participants described how environment might impact modality choice. P8 and P10 indicated that they would be hesitant to talk to Graphy while working in close proximity with co-workers, and suggested the need for headphones in these scenarios, while P6 said that in loud environments, they would “type commands, but at home [I] would talk to it.”

Prior experience with technology also played a role. This was strongest when it came to willingness to use touch gestures and Graphy’s conversational agent. Participants with extensive experience with smartphones appeared more willing to incorporate touch gestures. P6, who chose to zoom into the visualizations using a double tap and hold, discussed how their choice “corresponded with iPhones.” On the other hand, P8 described preferring interacting using voice, “I think I am so accustomed to my smart speakers and getting information that way.” Two participants who did not use touch gestures often, but were comfortable relying on conversation during interactions, stated, “I am used to Google Assistant” (P4) and “it is easiest to get the information I want… [by] asking a question” (P9).

Design Feedback. Participant feedback centered on labelling and complexity limits. Seven participants desired more extensive braille labelling. While braille titles and legends were provided, participants wanted the axes of the visualizations to be labelled. Six of them stated that they should be able to directly read braille labels or tap on them to have Graphy read the labels out aloud. Four participants (P1, P2, P6, and P11) acknowledged that due to the space and resolution constraints of the RTD, it may be difficult to sufficiently label the visualizations, with P2 suggesting that a touch gesture could summon/hide labels, “displaying axis markers takes up real estate, so having those not showing all the time, being able to bring them up easily… a press gesture… would be useful.” Another possibility, suggested by P2, was to render touch discernible symbols on the RTDs display, which users can tap to extract auditory labels.

Four participants (P1, P2, P6, and P8) felt that there was a limit on the complexity of visualizations that can be effectively understood on Graphy’s RTD. This related to the number of pins and pin heights that the RTD currently supports. Prior to being introduced to the Water line chart, participants were asked if they wanted to display water catchment storage, rainfall, or both data together. Only two participants (P1 and P3) chose to show them together when completing the analysis tasks, while three participants (P5, P8, and P10) speculated that displaying both datasets would prove overwhelming. Indeed, when asked to perform the adding data set operation, P8 reported difficulty interpreting sections of the multi-line chart where rainfall and water storage overlapped.

P8 discussed how the RTD’s pin heights were easy to read in isolation, e.g., in a six dot braille cell, but difficult to read in close proximity to one another, and P6 said that in close proximity, the pin heights did not have enough ‘contrast.’ P1 and P2 focused on the low pin resolution of the RTD, describing the isarithmic map as a ‘pixelated version’ of Australia. P2 also felt that Graphy was capable of rendering visualizations at lower resolutions than what they were used to with tactile graphics.

Keyboard input was also requested by three participants, to allow shortcuts, or as an alternative to speech input and touch gestures when manipulating visualizations, e.g., filtering or transforming.

4 Discussion and Limitations

4.1 The Combination of Conversational Agents and RTDs is Beneficial and Preferred

Five participants shared stories of frustration when being under-equipped to participate in data literacy activities during school years, highlighting the need for accessible data tools. This aligns with prior research showing that BLV students are underrepresented in STEM disciplines in tertiary education, due to unequal access [7].

Participants’ overall reaction toward the combination of RTD and conversational agent was very positive. They indicated that the combination would support both basic and more advanced data analysis. We observed that all participants used a mixture of touch, touch gestures, and natural language queries in their interactions with data and visualization. There was also a clear preference for the combination of RTD and conversational agent over a conversational agent, RTD, or printed tactile graphic alone.

To some extent, this preference for the combination of RTD and conversational agent is to be expected, given that multiple modalities increase adaptability [52]. In line with this, our participants indicated that different modalities would allow them to cater to personal preferences and the task context and purpose.

However, our results also revealed that prior experience with tactile graphics played a key role in the choice of modality. In particular, participants with lower tactile graphic experience leveraged the conversational agent to a greater degree. This aligns with earlier research by Reinders et al. [54], where participants with lower tactile experience used a conversational interface embedded in a multimodal interactive 3D printed model earlier and more frequently than participants with higher experience. Nonetheless, all participants, even those with little tactile graphic experience, strongly preferred the combination of RTD and conversational agent over a conversational agent alone. Furthermore, one participant (P8) stated that over time, they may be more willing to use touch more frequently and confidently, suggesting that less experienced BLV users may build tactile confidence and adjust their choice of interaction after using the system.

We were surprised by the strength of preference for tactile visualizations by users with more tactile experience. Participants spoke about how a tactile graphic allowed them to visualize the data, and most of those with greater tactile experience did not want to simply “be told the answer,” preferring to assemble their own understanding of the data through touch first. This desire for independent understanding came in spite of the additional time required for tactile exploration. Independence of interpretation has been recognized as a critical factor for increasing confidence in data-based decision-making by BLV people [16]. For experienced tactile graphics users, the RTD provided BLV users with the opportunity for individual sense-making, just as visual graphics do for sighted users, in a manner that is complementary to information provided by a conversational agent.

4.2 Is Graphy Practical?

A key question raised by any WOz study is whether the interface desired by participants can be constructed using current technologies. Because we catered for a broad range of queries pertaining to summary and background information about the visualizations (informed by [32]), almost all participant queries could be fulfilled using the scripts we prepared. The generation of these scripts was straightforward, and we believe could be readily supported by existing conversational agent platforms such as Dialogflow222https://cloud.google.com/dialogflow and Rasa333https://rasa.com/docs/rasa/.

One exception was the Fire isarithmic map, when participants were asked to determine the number of fire-weather days for a specific city. One participant (P10) asked for the location of the city and was given directions generated by the wizard on the fly. Another instance occurred when P10 was exploring the map and asked the agent if there had been a change in dangerous fire weather in a small regional city they had once visited while holidaying. Answering questions such as this requires the conversational agent to have an understanding of background information about the data. In principle, this may be supported by means of large language models (LLMs), but it raises issues of reliability and trust.

Participants also desired control over the flow of speech output, including the ability to pause, resume, or stop the conversational output. Placing users in control during conversational interactions has been identified as a critical need with mainstream conversational agents [1, 4]. Based on our experience designing conversational agents with BLV co-designers [53], this would be straightforward to support using Speech Synthesis Markup Language444https://www.w3.org/TR/speech-synthesis11/.

Our study also suggested that participants valued the use of pin height and single-touch and multi-touch gestures. Unfortunately, cheaper RTDs, such as the Dot Pad, do not support variable-height pins, and are not touch-sensitive. Even touch-sensitive RTDs, such as the Graphiti or Monarch, only support single-finger touch gestures and require the user to lift their other fingers when completing gestures. Thus, how best to support touch gestures remains an open issue.

4.3 How Do We Design Interactive Tactile Visualizations?

One area that needs further research before we can build an actual system, is a better understanding of how to design interactive tactile visualizations for an RTD. While there exists well-established guidance on visualization best practices when it comes to different types of data and tasks [42], such guidelines do not exist for the design of tactile visualizations for RTDs.

For instance, at present, we do not know the relative effectiveness of tactile line graphs and bar charts for different tasks, or the affordances they offer in this context. In our study, we observed that participants tended to prefer bar charts over line graphs. P6 described “the bar chart is easier to follow, I think it [allows me to] more accurately feel data at any particular point.” This seemed to be particularly true for participants with lower tactile experience, with P9 stating that “[with] the bar chart [it] was a lot easier to determine changes, [both the] highs and lows.” We observed that participants tended to trace line graphs using one or two fingers, with many tracing back and forth between inflection points, and that they traced up and down individual bars in a bar chart, and that in some cases, two hands were used to trace multiple bars. This hints that, just as for visual line graphs and bar charts [42], tactile line charts are suitable for exploration of trends, while bar charts are suitable for discrete comparisons. However, more research is required to test this.

While there are guidelines for the design of traditional tactile graphics (e.g., [44]), there are significant differences between these and tactile graphics displayed on RTDs. One important difference is the current low resolution of RTDs, and the consequent inability to use braille labels, e.g., markers and values on axes. Participants identified possible solutions, but more research is required.

Another area requiring research is how best to support zooming and panning. Touch gestures were used by six participants to pan (swipe) and zoom (pinch). However, when panning or zooming, five participants (P1, P2, P6, P7, and P8) became confused and lost their location when the RTD was refreshed. P1 said “Wow… that is crazy… I lost where I was.” P7 spoke of how “when zooming, [you need] to not lose your frame of reference, to be able to find it again.” P2 directly suggested that a scroll bar could be used to avoid this, “if you are panning a graph you could have an indicator where spatially in the graph you are, for example, you could have one column and row of pins, if you’re viewing the bottom then the column would be lower. Like a scroll bar would show where you are in a page.” Previous research with RTDs has suggested the use of buttons, rather than gestures, for operations like zooming to avoid context loss and user confusion [73, 71]. More research is needed into ways in which multimodal feedback can be used to retain user understanding of position and context during interactions such as panning and zooming, filtering, and search.

A further interaction warranting investigation is using pin actuation to highlight points or regions of interest on the RTD. Actuation has previously been used to highlight routes on maps [22]. We did not support actuation in our study but actuation could, for instance, support search and multimodal output where data points referenced in audio output are concurrently actuated in the visualization. This would be analogous to the use of speech with haptic vibratory feedback in interactive 3D models [53]. Such multimodal output would have been particularly useful with the Fire isarithmic map, when a participant asked for the location of a city and was given directions from the agent, which would have been more effective if combined with actuation.

4.4 Limitations

As the first step to understand the benefits of the combination of an RTD and conversational agent, our study involved only three visualization types – line graphs, bar charts, and isarithmic maps – and four task types – exploration, value identification, extremum identification, and trend. Future research needs to consider a broader range of visualization types and examine additional tasks users perform with these visualizations. We also primarily focused on the presentation of univariate spatial and time series data. Visualizations of more complex multivariate data, e.g., multiple time series visualizations, should be investigated.

As the majority of our participants were totally blind, we are also interested in recruiting additional legally blind participants, specifically those who still have residual vision, as their perspective may differ from that of people who are totally blind.

Our findings should also be interpreted within the limitations of the WOz method. While our findings strongly suggest that a combination of RTD and conversational agent is an effective tool for data analysis, it is crucial to implement and evaluate an actual system with BLV users. It would also be important to test the system with participants over a longer period of time and in real-world settings, e.g., in-the-wild studies in schools and workplaces.

5 Conclusion

Supporting BLV people’s data analysis remains an important and challenging question. Here, we investigated a novel approach that combines an RTD with a conversational agent by conducting a Wizard-of-Oz study with 11 BLV participants.

From the analysis of participants’ interactions, we identified a variety of multimodal interaction patterns. They depended on tasks and participants’ familiarity with tactile graphics. However, participants also made it clear that other factors such as purpose, location, and prior experience with smart phones and conversational agents influenced their interactions.

Our results show that all participants blended different modalities. Furthermore, participants suggested that the RTD and agent are complementary, and strongly believed that the multimodal combination supported autonomous data analysis and offered significant benefits over an RTD-only or a conversational-agent-only interface, or traditional tactile graphics.

We believe that the observations and findings, as well as participant suggestions from our study provide a solid basis for the design of a system supporting a combination of RTD and conversational agent, one that can allow BLV people to meaningfully engage independently in data access and analysis.

Acknowledgements.
We gratefully acknowledge the Australian Research Council’s Discovery Projects funding scheme (DP220101221). This work was supported in part by the Institute of Information and Communications Technology Planning and Evaluation (IITP) Grant funded by the Korean Government (MSIT), Artificial Intelligence Graduate School Program, Yonsei University, under Grant RS-2020-II201361.

References

  • [1] A. Abdolrahmani, R. Kuber, and S. M. Branham. "siri talks at you": An empirical investigation of voice-activated personal assistant (vapa) usage by individuals who are blind. In Proc. ACM SIGACCESS Conference on Computers & Accessibility, ASSETS ’18, pp. 249–258. ACM, New York, 2018. doi: 10 . 1145/3234695 . 3236344
  • [2] M. Z. I. Alam, S. Islam, and E. Hoque. Seechart: Enabling accessible visualizations through interactive natural language interface for people with visual impairments. In Proc. International Conference on Intelligent User Interfaces, IUI ’23, p. 46–64. ACM, New York, 2023. doi: 10 . 1145/3581641 . 3584099
  • [3] American printing house for the blind - meet monarch. https://www.aph.org/meet-monarch/, 2023.
  • [4] S. M. Branham and A. R. Mukkath Roy. Reading between the guidelines: How commercial voice assistant guidelines hinder accessibility for blind users. In Proc. ACM SIGACCESS Conference on Computers & Accessibility, ASSETS ’19, p. 446–458. ACM, New York, 2019. doi: 10 . 1145/3308561 . 3353797
  • [5] L. Brayda, F. Leo, C. Baccelliere, E. Ferrari, and C. Vigini. Updated tactile feedback with a pin array matrix helps blind people to reduce self-location errors. Micromachines, 14(9):351, 2018. doi: 10 . 3390/mi9070351
  • [6] L. Brayda, F. Leo, C. Baccelliere, C. Vigini, and E. Cocchi. A refreshable tactile display effectively supports cognitive mapping followed by orientation and mobility tasks: A comparative multi-modal study involving blind and low-vision participants. In Proc. Multimedia for Accessible Human Computer Interfaces, MAHCI ’19, 7 pages, p. 9–15. ACM, New York, 2019. doi: 10 . 1145/3347319 . 3356840
  • [7] M. Butler, L. Holloway, K. Marriott, and C. Goncu. Understanding the graphical challenges faced by vision-impaired students in australian universities. Higher Education Research & Development, 36(1):59–72, 2017. doi: 10 . 1080/07294360 . 2016 . 1177001
  • [8] M. Butler, L. M. Holloway, S. Reinders, C. Goncu, and K. Marriott. Technology developments in touch-based accessible graphics: A systematic review of research 2010-2020. In Proc. CHI Conference on Human Factors in Computing Systems, CHI ’21, article no. 278, 15 pages. ACM, New York, 2021. doi: 10 . 1145/3411764 . 3445207
  • [9] L. Cavazos Quero, J. Iranzo Bartolomé, D. Lee, Y. Lee, S. Lee, and J. Cho. Jido: A conversational tactile map for blind people. In Proc. ACM SIGACCESS Conference on Computers & Accessibility, ASSETS ’19, p. 682–684. ACM, New York, 2019. doi: 10 . 1145/3308561 . 3354600
  • [10] L. Cavazos Quero, J. Iranzo Bartolomé, S. Lee, E. Han, S. Kim, and J. Cho. An interactive multimodal guide to improve art accessibility for blind people. In Proc. ACM SIGACCESS Conference on Computers & Accessibility, ASSETS ’18, p. 346–348. ACM, New York, 2018. doi: 10 . 1145/3234695 . 3241033
  • [11] S. Connell, P.-Y. Kuo, L. Liu, and A. M. Piper. A wizard-of-oz elicitation study examining child-defined gestures with a whole-body interface. In Proc. Interaction Design and Children, IDC ’13, p. 277–280. ACM, New York, 2013. doi: 10 . 1145/2485760 . 2485823
  • [12] N. Dahlbäck, A. Jönsson, and L. Ahrenberg. Wizard of oz studies: why and how. In Proc. International Conference on Intelligent User Interfaces, IUI ’93, p. 193–200. ACM, New York, 1993. doi: 10 . 1145/169891 . 169968
  • [13] Dot inc - dot pad - the first tactile graphics display for the visually impaired. https://pad.dotincorp.com/, 2022.
  • [14] F. Elavsky, L. Nadolskis, and D. Moritz. Data navigator: An accessibility-centered data navigation toolkit. IEEE TVCG, 30(1):803–813, Jan 2024. doi: 10 . 1109/TVCG . 2023 . 3327393
  • [15] P. Ellis. Dual axes time series plots may be ok sometimes after all, 2016.
  • [16] D. Fan, A. Fay Siu, H. Rao, G. S.-H. Kim, X. Vazquez, L. Greco, S. O’Modhrain, and S. Follmer. The accessibility of data visualizations on the web for screen reader users: Practices and experiences during covid-19. ACM Trans. Access. Comput., 16(1), article no. 4, 29 pages, mar 2023. doi: 10 . 1145/3557899
  • [17] D. Fan, A. F. Siu, W.-S. A. Law, R. R. Zhen, S. O’Modhrain, and S. Follmer. Slide-tone and tilt-tone: 1-dof haptic techniques for conveying shape characteristics of graphs to blind users. In Proc. CHI Conference on Human Factors in Computing Systems, CHI ’22, article no. 477, 19 pages. ACM, New York, 2022. doi: 10 . 1145/3491102 . 3517790
  • [18] A. M. Farahani, P. Adibi, M. S. Ehsani, H.-P. Hutter, and A. Darvishy. Automatic chart understanding: A review. IEEE Access, 11:76202–76221, 2023. doi: 10 . 1109/ACCESS . 2023 . 3298050
  • [19] G. Fusco and V. S. Morash. The tactile graphics helper: Providing audio clarification for tactile graphics using machine vision. In Proc. ACM SIGACCESS Conference on Computers & Accessibility, ASSETS ’15, p. 97–106. ACM, New York, 2015. doi: 10 . 1145/2700648 . 2809868
  • [20] C. Goncu and K. Marriott. GraVVITAS: Generic multi-touch presentation of accessible graphics. In Human-Computer Interaction – INTERACT 2011, pp. 30–48. Springer, Berlin, Heidelberg, 2011. doi: 10 . 1007/978-3-642-23774-4_5
  • [21] S. Gyoshev, D. Karastoyanov, N. Stoimenov, V. Cantoni, L. Lombardi, and A. Setti. Exploiting a graphical braille display for art masterpieces. In K. Miesenberger and G. Kouroupetroglo, eds., Computers Helping People with Special Needs: 16th International Conference, ICCHP, pp. Part II, 237–245. Springer International Publishing, Switzerland, 2018. doi: 10 . 1007/978-3-319-94274-2_35
  • [22] L. Holloway, S. Ananthanarayan, M. Butler, M. T. De Silva, K. Ellis, C. Goncu, K. Stephens, and K. Marriott. Animations at your fingertips: Using a refreshable tactile display to convey motion graphics for people who are blind or have low vision. In Proc. ACM SIGACCESS Conference on Computers & Accessibility, ASSETS ’22, article no. 32, 16 pages. ACM, New York, 2022. doi: 10 . 1145/3517428 . 3544797
  • [23] L. Holloway, P. Cracknell, K. Stephens, M. Fanshawe, S. Reinders, K. Marriott, and M. Butle r. Refreshable tactile displays for accessible data visualisation. arXiv preprint arXiv:2401.15836, 2024. doi: 10 . 48550/arXiv . 2401 . 15836
  • [24] L. M. Holloway, C. Goncu, A. Ilsar, M. Butler, and K. Marriott. Infosonics: Accessible infographics for people who are blind using sonification and voice. In Proc. CHI Conference on Human Factors in Computing Systems, CHI ’22, article no. 480, 13 pages. ACM, New York, 2022. doi: 10 . 1145/3491102 . 3517465
  • [25] S. Holm. A simple sequentially rejective multiple test procedure. Scandinavian Journal of Statistics, 6(2):65–70, 1979.
  • [26] E. Hoque, P. Kavehzadeh, and A. Masry. Chart question answering: State of the art and future directions. Computer Graphics Forum, 41(3):555–572, 2022. doi: 10 . 1111/cgf . 14573
  • [27] J. Iranzo Bartolome, L. Cavazos Quero, S. Kim, M.-Y. Um, and J. Cho. Exploring art with a voice controlled multimodal guide for blind people. In Proc. Tangible, Embedded, and Embodied Interaction, TEI ’19, p. 383–390. ACM, New York, 2019. doi: 10 . 1145/3294109 . 3300994
  • [28] M. Ivanchev, F. Zinke, and U. Lucke. Pre-journey visualization of travel routes for the blind on refreshable interactive tactile displays. In K. Miesenberger, D. Fels, D. Archambault, P. Peňáz, and W. Zagler, eds., Computers Helping People with Special Needs, pp. 81–88. Springer International Publishing, Cham, 2014. doi: 10 . 1007/978-3-319-08599-9_13
  • [29] C. Jung, S. Mehta, A. Kulkarni, Y. Zhao, and Y.-S. Kim. Communicating visualizations without visuals: Investigation of visualization alternative text for people with visual impairments. IEEE TVCG, 28(1):1095–1105, Jan 2022. doi: 10 . 1109/TVCG . 2021 . 3114846
  • [30] P. H. Kahn, N. G. Freier, T. Kanda, H. Ishiguro, J. H. Ruckert, R. L. Severson, and S. K. Kane. Design patterns for sociality in human-robot interaction. In Proc. ACM/IEEE International Conference on Human Robot Interaction, HRI ’08, pp. 97–104. ACM, New York, 2008. doi: 10 . 1145/1349822 . 1349836
  • [31] J. F. Kelley. An iterative design methodology for user-friendly natural language office information applications. ACM Trans. Inf. Syst., 2(1):26–41, 16 pages, Jan. 1984. doi: 10 . 1145/357417 . 357420
  • [32] J. Kim, A. Srinivasan, N. W. Kim, and Y.-S. Kim. Exploring chart question answering for blind and low vision users. In Proc. CHI Conference on Human Factors in Computing Systems, CHI ’23, article no. 828, 15 pages. ACM, New York, 2023. doi: 10 . 1145/3544548 . 3581532
  • [33] N. W. Kim, S. C. Joyner, A. Riegelhuth, and Y. Kim. Accessible visualization: Design space, opportunities, and challenges. Computer Graphics Forum, 40(3):173–188, 2021. doi: 10 . 1111/cgf . 14298
  • [34] S. Kim, Y. Ryu, J. Cho, and E.-S. Ryu. Towards tangible vision for the visually impaired through 2d multiarray braille display. Sensors, 19(23), 2019. doi: 10 . 3390/s19235319
  • [35] G. Kramer, B. Walker, T. Bonebright, P. Cook, J. H. Flowers, N. Miner, and J. Neuhoff. Sonification report: Status of the field and research agenda. Technical report, Department of Psychology, University of Nebraska, Lincoln, Nebraska, 2010.
  • [36] B. Lee, E. K. Choe, P. Isenberg, K. Marriott, and J. Stasko. Reaching broader audiences with data visualization. IEEE computer graphics and applications, 40(2):82–90, 2020. doi: 10 . 1109/MCG . 2020 . 2968244
  • [37] B. Lee, K. Marriott, D. Szafir, and G. Weber. Inclusive Data Visualization (Dagstuhl Seminar 23252), 2024. doi: 10 . 4230/DagRep . 13 . 6 . 81
  • [38] K. Marriott, B. Lee, M. Butler, E. Cutrell, K. Ellis, C. Goncu, M. Hearst, K. McCoy, and D. A. Szafir. Inclusive data visualization for people with disabilities: a call to action. Interactions, 28(3):47–51, Apr 2021. doi: 10 . 1145/3457875
  • [39] metec ag - two-dimensional, touch-sensitive graphic displays. https://metec-ag.de/en/produkte-graphik-display.php, 2012.
  • [40] J. A. Miele, S. Landau, and D. Gilden. Talking tmap: Automated generation of audio-tactile maps using smith-kettlewell’s tmap software. British Journal of Visual Impairment, 24(2):93–100, 2006. doi: 10 . 1177/0264619606064436
  • [41] T. Motoyoshi, S. Mizushima, K. Sawai, T. Tamamoto, H. Masuta, K. Koyanagi, and T. Oshima. Prototype development of a shape presentation system using linear actuators. In ICCHP: International Conference on Computers Helping People with Special Needs, pp. Part II, 226–230. Springer International, Switzerland, 2018. doi: 10 . 1007/978-3-319-94274-2_31
  • [42] T. Munzner. Visualization Analysis and Design. AK Peters Visualization Series. CRC Press, 2015. doi: 10 . 1201/b17511
  • [43] R. K. Namdev and P. Maes. An interactive and intuitive stem accessibility system for the blind and visually impaired. In PETRA: International Conference on PErvasive Technologies Related to Assistive Environments, pp. 1–7. ACM, 2015. doi: 10 . 1145/2769493 . 2769502
  • [44] B. A. of North America. Guidelines and Standards for Tactile Graphics. Braille Authority of North America, 2010.
  • [45] H. Ohshima, M. Kobayashi, and S. Shimada. Development of blind football play-by-play system for visually impaired spectators: Tangible sports. In Extended Abstracts of the CHI Conference on Human Factors in Computing Systems, CHI EA ’21, article no. 209, 6 pages. ACM, New York, 2021. doi: 10 . 1145/3411763 . 3451737
  • [46] Orbit research - graphiti - a breakthrough in non-visual access to all forms of graphical information. http://www.orbitresearch.com/product/graphiti/, 2016.
  • [47] S. O’Modhrain, N. A. Giudice, J. A. Gardner, and G. E. Legge. Designing media for visually-impaired users of refreshable touch displays: Possibilities and pitfalls. IEEE ToH, 8(3):248–257, 2015. doi: 10 . 1109/TOH . 2015 . 2466231
  • [48] T. Park, J. Jung, and J. Cho. A method for automatically translating print books into electronic braille books. Science China Information Sciences, 59, 2016. doi: 10 . 1007/s11432-016-5575-z
  • [49] A. Preece. Spotlighting breakthroughs in multi-line braille displays. https://www.afb.org/blog/entry/spotlighting-multi-line-displays, 2024.
  • [50] R. Puduppully, L. Dong, and M. Lapata. Data-to-text generation with entity modeling. In Proc. Association for Computational Linguistics, pp. 2023–2035. ACL, Florence, Italy, 2019. doi: 10 . 18653/v1/P19-1195
  • [51] C. Rebuffel, L. Soulier, G. Scoutheeten, and P. Gallinari. A hierarchical model for data-to-text generation. In J. M. Jose, E. Yilmaz, J. Magalhães, P. Castells, N. Ferro, M. J. Silva, and F. Martins, eds., Advances in Information Retrieval, pp. 65–80. Springer, Cham, 2020. doi: 10 . 1007/978-3-030-45439-5_5
  • [52] L. M. Reeves, J. Lai, J. A. Larson, S. Oviatt, T. S. Balaji, S. Buisine, P. Collings, P. Cohen, B. Kraal, J.-C. Martin, M. McTear, T. Raman, K. M. Stanney, H. Su, and Q. Y. Wang. Guidelines for multimodal user interface design. CACM, 47(1):57–59, Jan. 2004. doi: 10 . 1145/962081 . 962106
  • [53] S. Reinders, S. Ananthanarayan, M. Butler, and K. Marriott. Designing conversational multimodal 3d printed models with people who are blind. In Proc. ACM Designing Interactive Systems Conference, DIS ’23, 17 pages, p. 2172–2188. ACM, New York, 2023. doi: 10 . 1145/3563657 . 3595989
  • [54] S. Reinders, M. Butler, and K. Marriott. "hey model!" – natural user interactions and agency in accessible interactive 3d models. In Proc. CHI Conference on Human Factors in Computing Systems, CHI ’20, 13 pages, p. 1–13. ACM, New York, 2020. doi: 10 . 1145/3313831 . 3376145
  • [55] E. M. Rogers. Diffusion of innovations. Free Press, New York, 5th ed., 08 2003.
  • [56] J. Rowell and S. Ungar. The world of touch: an international survey of tactile maps. part 1: production. British Journal of Visual Impairment, 21(3):98–104, 2003. doi: 10 . 1177/02646196030210030
  • [57] D. Salber and J. Coutaz. Applying the wizard of oz technique to the study of multimodal systems. In L. J. Bass, J. Gornostaev, and C. Unger, eds., Human-Computer Interaction, pp. 219–230. Springer Berlin Heidelberg, Berlin, Heidelberg, 1993. doi: 10 . 1007/3-540-57433-6_51
  • [58] SAS graphics accelerator customer product page. https://support.sas.com/software/products/graphics-accelerator.
  • [59] B. Schmitz and T. Ertl. Interactively displaying maps on a tactile graphics display. In SKALID 2012 Spatial Knowledge Acquisition with Limited Information Displays, pp. 13–18, 2012.
  • [60] J. Seo, Y. Xia, B. Lee, S. Mccurry, and Y. J. Yam. Maidr: Making statistical visualizations accessible with multimodal data representation. In Proc. CHI Conference on Human Factors in Computing Systems, CHI ’24, article no. 211, 22 pages. ACM, New York, 2024. doi: 10 . 1145/3613904 . 3642730
  • [61] A. Sharif, O. H. Wang, A. T. Muongchan, K. Reinecke, and J. O. Wobbrock. VoxLens: Making online data visualizations accessible with an interactive javascript plug-in. In Proc. CHI Conference on Human Factors in Computing Systems, CHI ’22, article no. 478, 19 pages. ACM, New York, 2022. doi: 10 . 1145/3491102 . 3517431
  • [62] L. Shi, H. Lawson, Z. Zhang, and S. Azenkot. Designing interactive 3d printed models with teachers of the visually impaired. In Proc. CHI Conference on Human Factors in Computing Systems, CHI ’19, article no. 197, 14 pages. ACM, New York, 2019. doi: 10 . 1145/3290605 . 3300427
  • [63] L. Shi, Y. Zhao, and S. Azenkot. Designing interactions for 3d printed models with blind people. In Proc. ACM SIGACCESS Conference on Computers & Accessibility, ASSETS ’17, pp. 200–209. ACM, New York, 2017. doi: 10 . 1145/3132525 . 3132549
  • [64] A. Siu, G. S-H Kim, S. O’Modhrain, and S. Follmer. Supporting accessible data visualization through audio data narratives. In Proc. CHI Conference on Human Factors in Computing Systems, CHI ’22, article no. 476, 19 pages. ACM, New York, 2022. doi: 10 . 1145/3491102 . 3517678
  • [65] J. R. Thompson, J. J. Martinez, A. Sarikaya, E. Cutrell, and B. Lee. Chart reader: Accessible visualization experiences designed with screen reader users. In Proc. CHI Conference on Human Factors in Computing Systems, CHI ’23, article no. 802, 18 pages. ACM, New York, 2023. doi: 10 . 1145/3544548 . 3581186
  • [66] F. Vidal-Verdu and M. Hafez. Graphical tactile displays for visually-impaired people. IEEE Transactions on Neural Systems and Rehabilitation Engineering, 15(1):119–130, 2007. doi: 10 . 1109/TNSRE . 2007 . 891375
  • [67] J. Walny, B. Lee, P. Johns, N. Henry Riche, and S. Carpendale. Understanding pen and touch interaction for data exploration on interactive whiteboards. IEEE TVCG, 18(12):2779–2788, 2012. doi: 10 . 1109/TVCG . 2012 . 275
  • [68] J. O. Wobbrock, M. R. Morris, and A. D. Wilson. User-defined gestures for surface computing. In Proc. CHI Conference on Human Factors in Computing Systems, CHI ’09, pp. 1083–1092. ACM, New York, 2009. doi: 10 . 1145/1518701 . 1518866
  • [69] W. Yang, J. Huang, R. Wang, W. Zhang, H. Liu, and J. Xiao. A survey on tactile displays for visually impaired people. IEEE Transactions on Haptics, 14(4):712–721, 2021. doi: 10 . 1109/TOH . 2021 . 3085915
  • [70] J. Yu, E. Reiter, J. Hunter, and C. Mellish. Choosing the content of textual summaries of large time-series data sets. Natural Language Engineering, 13(1):25–49, 2007. doi: 10 . 1017/S1351324905004031
  • [71] L. Zeng, M. Miao, and G. Weber. Interactive audio-haptic map explorer on a tactile display. Interacting with Computers, 27(4):413–429, 2015. doi: 10 . 1093/iwc/iwu006
  • [72] L. Zeng, D. Prescher, and G. Weber. Exploration and avoidance of surrounding obstacles for the visually impaired. In Proc. ACM SIGACCESS Conference on Computers & Accessibility, ASSETS ’12, p. 111–118. ACM, New York, 2012. doi: 10 . 1145/2384916 . 2384936
  • [73] L. Zeng, G. Weber, I. Zoller, P. Lotz, T. A. Kern, J. Reisinger, T. Meiss, T. Opitz, T. Rossner, and N. Stefanova. Examples of haptic system development, pp. 525–554. Springer, 2014. doi: 10 . 1007/978-3-031-04536-3_14
  • [74] Z. Zhang, J. R. Thompson, A. Shah, M. Agrawal, A. Sarikaya, J. O. Wobbrock, E. Cutrell, and B. Lee. ChartA11y: Designing accessible touch experiences of visualizations with blind smartphone users. In Proc. ACM SIGACCESS Conference on Computers & Accessibility, ASSETS ’24. ACM, New York, 2024. in press.