This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

An Interactive Tool for Simulating Mid-Air Ultrasound Tactons on the Skin

Chungman Lim [email protected] 0000-0002-7857-3322 Gwangju Institute of
Science and Technology
Republic of Korea
Hasti Seifi [email protected] 0000-0001-6437-0463 Arizona State UniversityTempeUnited States  and  Gunhyuk Park [email protected] 0000-0003-2677-5907 Gwangju Institute of
Science and Technology
Republic of Korea
(2024)
Abstract.

Mid-air ultrasound haptic technology offers a myriad of temporal and spatial parameters for contactless haptic design. Yet, predicting how these parameters interact to render an ultrasound signal is difficult before testing them on a mid-air ultrasound haptic device. Thus, haptic designers often use a trial-and-error process with different parameter combinations to obtain desired tactile patterns (i.e., Tactons) for user applications. We propose an interactive tool with five temporal and three spatiotemporal design parameters that can simulate the temporal and spectral properties of stimulation at specific skin points. As a preliminary verification, we measured vibrations induced from the ultrasound Tactons varying on one temporal and two spatiotemporal parameters. The measurements and simulation showed similar results for three different ultrasound rendering techniques, suggesting the efficacy of the simulation tool. We present key insights from the simulation and discuss future directions for enhancing the capabilities of simulations.

Mid-Air Haptics, Computational Simulation, Ultrasound Tacton Design
journalyear: 2024copyright: rightsretainedconference: Extended Abstracts of the CHI Conference on Human Factors in Computing Systems; May 11–16, 2024; Honolulu, HI, USAbooktitle: Extended Abstracts of the CHI Conference on Human Factors in Computing Systems (CHI EA ’24), May 11–16, 2024, Honolulu, HI, USAdoi: 10.1145/3613905.3650981isbn: 979-8-4007-0331-7/24/05ccs: Human-centered computing User interface toolkitsccs: Human-centered computing HCI theory, concepts and models
Refer to caption
Figure 1. Overview of our interactive simulation tool for testing design parameters in mid-air ultrasound technology. Haptic designers can control five temporal and three spatiotemporal parameters to design a mid-air ultrasound Tacton. They can also select a point on the 2D plane to visualize the temporal waveform and frequency spectrum of the Tacton at that point on the skin.
\Description

1. Introduction

Mid-air ultrasound technologies create haptic feedback on the user’s skin without physical contact with a device. This technology focuses acoustic waves into one or multiple focal points using a phased array of ultrasonic transducers, and modulates or moves these focal points to create a sense of touch (Rakkolainen et al., 2020). Designers are exploring the parameter space of the mid-air ultrasound patterns (i.e., tactile icons or Tactons) to deliver information or emotion to users in various applications, including touchless public displays (Limerick et al., 2019; Vi et al., 2017), automotive user interfaces (Harrington et al., 2018; Brown et al., 2022), medical training simulations (Hung et al., 2013, 2014), and virtual reality environments (Howard et al., 2022; Hwang et al., 2017b; Mulot et al., 2023b; Villa et al., 2022).

Several rendering techniques and parameters exist for creating mid-air ultrasound Tactons. The common approaches include using either temporal parameters (e.g., amplitude-modulated frequency) (Hoshi et al., 2009), spatiotemporal parameters (e.g., trajectory or drawing speed of a focal point) (Frier et al., 2018; Takahashi et al., 2019; Mulot et al., 2023a), or a combination of both (Hajas et al., 2020; Rutten et al., 2020; Dalsgaard et al., 2022). These rendering techniques accompany complex physical effects on the skin. For example, amplitude modulation (AM) focuses acoustic pressures on a static focal point, vibrating the local distribution of skin and activating a group of mechanoreceptors. Spatiotemporal modulation (STM) moves a focal point along a trajectory, thus vibrating the skin at a drawing frequency on the trajectory points. Furthermore, the combination of AM and STM techniques offers both AM frequency and drawing frequency, making it challenging for haptic designers to predict the rendered frequency on the user’s palm.

The combinations of parameters from the above techniques usually yield in vibrations that are complex to predict, so the designers resort to repeat trial-and-errors to find a Tacton set of their interests. Thus, the physical simulations of the mid-air ultrasound Tacton can help understand its perception and reduce the Tacton design cost. Prior literature proposes physical simulations for the ultrasound vibrations and their measurement data, for example, AM or STM of a focal point (Carter et al., 2013; Chilles et al., 2019) and gap detection thresholds between two static focal points (Carter et al., 2013; Howard et al., 2023). Yet, as the existing simulations typically simulate one parameter or one rendering technique, more research is needed on interactive simulation tools for testing the interaction of temporal and spatiotemporal parameters in complex ultrasound Tactons and the combination of rendering techniques.

To fill this gap, we developed a Python-based interactive simulation tool for skin vibrations induced by ultrasound Tactons rendered by modulating a single focal point. Our simulation facilitates the design of Tactons with five temporal parameters: amplitude, AM frequency, envelope frequency, superposition ratio, and total duration; and three spatiotemporal parameters: the shape, size, and drawing speed of a focal point’s trajectory. These parameters are commonly used by designers for creating Tactons. Designers can manipulate these eight parameters in the control panel and view the vibration waveform and frequency spectrum at any point on the skin in the visualization panel. Thus, the simulation tool enables designers to efficiently explore the physical effects of multiple parameters before rendering the Tactons on the device.

As an initial verification of the simulation tool, we designed 15 mid-air ultrasound Tactons varying in AM frequency, size, and drawing speed and tested AM rendering, STM rendering, and a combination of AM and STM rendering techniques. We selected the three parameters in our preliminary measurements because they mainly affect the spectral peaks of the induced vibrations at a skin point. We used our simulation tool to predict vibrations at five points on the skin. Then, we employed a STRATOS Explore ultrasound haptic device to render the Tactons and measured vibrations induced at these points with a laser vibrometer. Our preliminary measurements showed high correspondence with our simulation results, revealing similar temporal waveforms and spectral harmonics to the simulated predictions. We discuss directions for future research to improve the measurement methodology and to build a high-utility simulation. Our contributions include:

  • An interactive tool that simulates the vibrations induced on the skin from physical interactions of five temporal and three spatiotemporal parameters in mid-air ultrasound Tactons.

  • Preliminary measurements of vibrations induced by 15 mid-air ultrasound Tactons, suggesting the validity of the simulation tool for AM and STM rendering.

2. Method and implementation

We developed a computational model to simulate mid-air ultrasound Tactons that use a single focal point. This model predicts the vibration waveforms and frequency spectra at any points on the skin area.

2.1. Assumptions for Simulation

We made several assumptions about mid-air ultrasound stimulation in order to lower the computational complexity of the simulation tool. In the literature, focused mid-air ultrasound at a single focal point creates a central oval-shaped vibration area with maximum amplitude, followed by four smaller vibration areas located at four directions apart 15 cmcm from the center with about less than 30 percent of the center amplitude (side lobes) if the ultrasound device located at the 20 cmcm distance (Carter et al., 2013; Wilson et al., 2014). The amplitudes of side lobes are nearly below the detection threshold at the maximum stimulation (Howard et al., 2019), while the side lobes vary on the distance between the device and the focal point due to changes in the angles of the ultrasound transducers. Therefore, our model assumes the ultrasound stimulation occurs at a target single circular point on the skin. In addition, we assumed that the skin of the user’s palm is a 2D plane, and the propagation of waves along the skin does not occur, considering the computational complexity of the model. We discuss the implications of these assumptions in Section 4.

2.2. Parameter Space of Mid-Air Ultrasound Tactons in Simulation

Refer to caption
Refer to caption
(a) Intensity by a static focal point
(b) Relative intensity by height from a device
Refer to caption
Refer to caption
(c) Control panel
(d) Visualization panel
Figure 2. Plots for the intensity of stimulation by a single focal point at the height h and the interactive simulation tool: (a) The intensity of stimulation decreases with distance from the focal point (Carter et al., 2013). (b) The relative intensity determined by the distance between a focal point and the mid-air ultrasound device. (c) The control panel in our interactive tool for manipulating parameters in temporal and spatiotemporal configurations. The dropdowns allow users to select design parameters and a position (a=(a,b)\vec{a}=(a,~{}b)) on the skin. (d) The visualization panel showing the 2D stimulated skin area (Left) and vibration waveform (Right) at a specific point on the skin. The sky-colored line represents the trajectory of a focal point, the black dotted line represents the borderline influenced by the stimulation, and the “X” symbol represents the point selected to see the effects of the stimulation by the Tacton. The three plots (Right) display the temporal plot of the Tacton, and the temporal and spectral plots of the stimulation at the selected point a\vec{a} by the user .
Refer to caption
Refer to caption
(a)
(b)
Figure 3. Our measurement setup. (a) We used a laser vibrometer to measure vibrations on paper induced by mid-air ultrasound Tactons. (b) In all measurements, we measured the vibrations at the same 5 points, spaced at 5 mm intervals.

Our model provides two configuration spaces for rendering mid-air ultrasound Tactons (Figure 1): (1) The temporal configuration includes five parameters of amplitude, AM frequency, envelope frequency, superposition ratio, and total duration; and (2) the spatiotemporal configuration includes three parameters of shape, size, drawing speed of a focal point trajectory. We selected these parameters based on the most frequently used parameters in Tacton design literature (Dalsgaard et al., 2022; Hajas et al., 2020; Obrist et al., 2015, 2013; Hwang et al., 2017a; Seifi et al., 2015; Park and Choi, 2011). Thus, users can observe the vibration waveform at any point on the skin by controlling these eight parameters of mid-air ultrasound Tactons.

1. Temporal configuration: We defined the temporal configuration of mid-air ultrasound Tactons using the mathematical equation p(t)p(t), which controls five parameters:

(1) p(t)={AU(t){wAM1M1(t)+wAM2M2(t)}E(t),if superposition ratio is used.AU(t)M(t)E(t),otherwise.p(t)=\begin{dcases}A\cdot U(t)\{w_{AM_{1}}M_{1}(t)+w_{AM_{2}}M_{2}(t)\}E(t),&\!\!\!\text{if {superposition ratio} is used.}\\ A\cdot U(t)M(t)E(t),&\text{otherwise.}\end{dcases}

where U(t)U(t) represents the continuous ultrasound at 40 kHz (typical frequency from commercial mid-air ultrasound haptic devices), M(t)M(t) is an AM sinusoid of sin(2πfAMt)sin(2\pi f_{AM}t) with AM frequency fAMf_{AM}, E(t)E(t) is an envelope sinusoid of sin(2πfet)sin(2\pi f_{e}t) with envelope frequency fef_{e}, and wAM1w_{AM_{1}} and wAM2w_{AM_{2}} is the weights of two AM sinusoids (i.e., M1(t)M_{1}(t) and M2(t)M_{2}(t)).

Amplitude (AA) corresponds to the peak acoustic pressure at the focal point, ranging between 0% and ±\pm 100%, as provided by a mid-air ultrasound device. While AA represents a commanded amplitude on the device, its relative intensity varies with height, which is the vertical distance between the focal point and the mid-air ultrasound device. The amplitude is known to have an inverted U-shaped relationship with height, reaching a maximum at 200 mmmm (Raza et al., 2019) above the device. We applied the above findings to calculate relative intensity of AA as in Figure 2(b). AM frequency (fAMf_{AM}) represents a temporal frequency that modulates U(t)U(t) (Obrist et al., 2013, 2015). Here, fAMf_{AM} = 0 Hz indicates no AM rendering (i.e., M(t)=1M(t)=1), suggesting that STM rendering is necessary to create tactile sensations. Envelope frequency (fef_{e}) refers to a frequency modulating M(t)M(t) (Park and Choi, 2011) and fef_{e} = 0 Hz denotes a constant envelope (i.e., E(t)=1E(t)=1). Superposition ratio (wAM1w_{AM_{1}}:wAM2w_{AM_{2}}) is the mixing ratio of two sinusoidal signals for creating a superimposed signal. The simulation tool currently provides five ratios: 1:0, 0.75:0.25, 0.5:0.5, 0.25:0.75, 0:1 (Yoo et al., 2022; Hwang et al., 2017a; Lim and Park, 2023). Total duration (tdt_{d}) is the maximum tt of the vibrations. While the computational model can handle any duration, we set 10 seconds as the maximum in the current tool, in line with the Tacton durations common in user applications (Seifi et al., 2015).

2. Spatiotemporal configuration: We defined the spatiotemporal configuration as the spatial properties of the temporal formula p(t)p(t), consisting of three parameters, shape, size, and drawing speed of a focal point trajectory in a 2D plane above the mid-air ultrasound device (Figure 1). In this configuration, shape represents the trajectory of a focal point in the 2D plane. Based on the literature (Rutten et al., 2019; Hajas et al., 2020), we provide five shapes: point, horizontal line, circle, regular triangle, and square. Size (dd, in mmmm) refers to the length of the line, diameter of the circle, or the side length for the regular triangle and square. We set the maximum size as 60 mmmm, considering the typical size of a user’s palm (Shen et al., 2023; Hajas et al., 2020). Drawing speed (vv, in m/sm/s) denotes the velocity of a focal point’s movement. Combinations of these three parameters derive a drawing frequency (fdf_{d}), defined as the number of completions or revolutions of a trajectory per second (Freeman and Wilson, 2021; Wojna et al., 2023; Rutten et al., 2020), creating a spatiotemporal tactile sensation (i.e., STM) (Frier et al., 2018). The three parameters of shape, size, and drawing speed determine the 2D position of a focal point on the moving trajectory at tt, represented as x(t)\vec{x}(t) = (x(t)x(t)y(t)y(t)). For example, when shape is the point, x(t)\vec{x}(t) is a constant at (0, 0). When shape is the circle, x(t)\vec{x}(t) can be expressed as (d2cos(2πfdt),d2sin(2πfdt)\frac{d}{2}\cos(2\pi f_{d}t),~{}\frac{d}{2}\sin(2\pi f_{d}t)). With the trajectory x(t)\vec{x}(t) from a total of the three parameters, we denoted p(t)p(t) at the focal point as p(x(t),t)p(\vec{x}(t),t).

Refer to caption
(a) PureAMPureAM
(fAMf_{AM} = 140 Hz)
Refer to caption
(b) PureSTMPureSTM (dd = 20 mmmm, vv = 12 m/sm/s, fdf_{d} = 191 Hz)
Refer to caption
(c) AM+STMAM{+}STM (fAMf_{AM} = 140 Hz, dd = 20 mmmm, vv = 12 m/sm/s, fdf_{d} = 191 Hz)
Figure 4. Three exemplar comparisons for pure AM rendering, pure STM rendering, and combination of AM and STM rendering, between simulations (blue) and measurements (red).

2.3. Model Architecture

At a selected point a=(a,b)\vec{a}=(a,b) on the 2D plane above the device, we can define the Euclidean distance between x(t)\vec{x}(t) (location of the focal point) and a\vec{a} as D(x(t),a)D(\vec{x}(t),\vec{a}). The distance determines the intensity of stimulation at a\vec{a}, which we defined as S(D(x(t),a))S(D(\vec{x}(t),\vec{a})) (Figure 2(a)). Thus, we defined the final intensity of stimulation as A(x(t),a)=AS(D(x(t),a))A(\vec{x}(t),\vec{a})=A\cdot S(D(\vec{x}(t),\vec{a})) and expressed p(x(t),t)p(\vec{x}(t),t) at a\vec{a} as:

(2) p(x(t),a,t)={A(x(t),a)U(t){wAM1M1(t)+wAM2M2(t)}E(t),if superposition ratio is used.A(x(t),a)U(t)M(t)E(t),otherwise.p(\vec{x}(t),\vec{a},t){=}\begin{dcases}\!A(\vec{x}(t),\vec{a})U(t)\{w_{AM_{1}}M_{1}(t){+}w_{AM_{2}}M_{2}(t)\}E(t),&\!\!\!\!\text{if {superposition ratio} is used.}\\ A(\vec{x}(t),\vec{a})U(t)M(t)E(t),&\text{otherwise.}\end{dcases}

With this implementation, our model facilitates tests for the temporal and spatiotemporal configurations, comprising five and three parameters, respectively (Figure 2(c)). In the interactive simulation tool, we provide the stimulated areas on the skin by p(x(t),t)p(\vec{x}(t),t) (Figure 2(d) Left). Our tool also presents the temporal plot of the commanded Tacton (i.e., Equation 1) considering U(t)U(t) as 40 kHzkHz sinusoid. For the temporal plot for the signal at a\vec{a} (i.e., Equation 2), we substituted 1 to U(t)U(t) because the ultrasound stimulation acts as a constant pressure (Frier et al., 2018; Shen et al., 2023). Then we applied Fourier transform to the temporal waveform to estimate the frequency spectrum (Figure 2(d) Right).

3. Preliminary Measurement

For preliminary verification for the capability of our simulation tool, we designed and measured 15 mid-air ultrasound Tactons considering three rendering scenario: (1) Pure AM rendering (PureAMPureAM), (2) pure STM rendering (PureSTMPureSTM), and (3) a combination of AM and STM (AM+STMAM{+}STM).

3.1. Methods

We selected ultrasound Tactons for the measurements, focusing on AM frequency (fAMf_{AM}) in the temporal configuration and size (dd) and drawing speed (vv) in the spatiotemporal configuration, as these three parameters mainly determine the AM frequency and the drawing frequency of the ultrasound Tactons which affects vibration spectrum. We did not use the other temporal parameters and we kept total duration at 1 second. In addition, we maintained shape as the circle and height at 200 mmmm. We used four fAMf_{AM} values: 0, 80, 140, and 210 Hz. We selected fAMf_{AM} = 0 Hz to test pure STM rendering (PureSTMPureSTM) and fAMf_{AM} ¿ 60 Hz, as the power lines in our country introduces a constant 60 Hz measurement noise. We also used four combinations of size and drawing speed as (dd in mmmm, vv in m/sm/s): (0, 0), (10, 6), (20, 12), (30, 18). We chose (0 mmmm, 0 m/sm/s) to test pure AM rendering (PureAMPureAM), and maintained the same ratio of vd\frac{v}{d} for the other three combinations to have the same drawing frequency at 191 Hz.

We rendered the ultrasound Tactons using the STRATOS Explore device by Ultraleap and measured the Tactons using a 1D laser vibrometer (Ploytec IVS-500) on paper (Figure 3(a)). Initially, we tried to measure the displacements induced by the ultrasound device on the skin of palm but the induced displacement was too weak and lower than the vibrometer’s minimum resolution. After much experimentation, we configured a system to measure displacements induced by the ultrasound Tactons on a thin paper. For the Tactons varying on AM frequency, size, and drawing speed, we measured vibrations at 5 points inside, on the trajectory, and outside the shape, spaced at 5 mm intervals (Figure 3(b)).

3.2. Results

The preliminary measurement results showed similar results to those reported in (Chilles et al., 2019), although we measured the vibrations on paper. PureAMPureAM introduced frequency harmonics (multiples of AM frequency or fAMf_{AM}) (Figure 4). PureSTMPureSTM rendering showed frequency harmonics (multiples of drawing frequency or fdf_{d}), which appeared consistently for different drawing speed and size as long as the drawing frequency (the ratio of speed to size) was the same. The combinations of AM and STM (AM+STMAM{+}STM) also resulted in frequency harmonics at multiples of both AM frequency and drawing frequency. Also, the highest magnitude occurred at the inputted AM frequency among all harmonic frequencies, regardless of the STM parameters.

Our model for PureAMPureAM and PureSTMPureSTM simulated the exact AM frequency and drawing frequency (fAMf_{AM} and fdf_{d}) as observed in the measurements (Figure 4). In particular, simulations for PureSTMPureSTM showed the same frequency harmonics at multiples of the drawing frequency. However, for both PureAMPureAM and AM+STMAM{+}STM, the simulations and measurements showed less correspondence, perhaps due to the different characteristics between paper and human skin. The simulated waveform for PureAMPureAM included a single component at the inputted fAMf_{AM} in the spectral domain, while the measured waveform for PureAMPureAM introduced frequency harmonics at multiples of fAMf_{AM} in the measurement. In AM+STMAM{+}STM, both the simulations and measurements showed the highest magnitude at the inputted AM frequency, regardless of the STM parameters. However, the simulations showed the fAMf_{AM} component and pair frequency components at multiples of fd±fAMf_{d}\pm f_{AM}, while the measurements included frequency harmonics at multiples of both fAMf_{AM} and fdf_{d}. These frequency harmonics have a much lower magnitude than the AM frequency, so the extent of their impact on user perception is not fully known.

4. Discussion

Our simulation can provide insights for designing mid-air ultrasound Tactons, as the designer can visualize the complex frequency spectrum induced by PureSTMPureSTM or AM+STMAM{+}STM and anticipate its impact on the end-user experience of the Tactons. In other words, our proposed simulation can inform actual stimulation at any points above the device, aiding as an effective means for testing the physical effects of the created Tactons by haptic designers.

The preliminary measurement data suggested the limitations of the current measurement setup and simulation model. At the current setup, the paper had different characteristics from human skin, such as elasticity and shear wave propagation, thus this measurement did not perfectly reflect the stimulation process on the human skin. Moreover, our model architecture was built on the simplified assumptions (Section 2.1) and on data collected from a scale hung in the air (Raza et al., 2019) and a microphone placed away from the device (Carter et al., 2013), instead of using human skin stimulation data.

In the future, we aim to improve the measurement methodology, for example, by using a higher resolution laser vibrometer or employing a silicon-based replica of human skin (Frier et al., 2018; Chilles et al., 2019). Also, we plan to improve the simulation’s accuracy by relaxing our current assumption such as the propagation of vibration waves on the skin to fully capture the complexity of mid-air ultrasound stimulation. Finally, we plan to improve our interactive simulation tool by including more design parameters, such as rhythmic structure and the number of focal points, and by adding the borderline of perceptible intensities considering detection threshold. We plan to make our simulation tool open-source after making the above improvements and further validating the model with a larger set of Tactons.

5. Conclusion

A physical simulation for skin vibrations can offer new possibilities for rapidly prototyping mid-air ultrasound Tactons for user applications. We proposed an interactive simulation for mid-air ultrasound Tactons that allows designers to easily test combinations of eight ultrasound parameters (five temporal and three spatiotemporal) and visualize the vibrations induced at different points on the skin. Our initial results suggest high correspondence between our simulation and measurements of a set of Tactons. We hope the simulation can assist haptic designers in creating rich mid-air ultrasound Tactons that vary on temporal and spatial parameters.

Acknowledgements.
We would like to thank Kyuyoung Shim and Gyungmin Jin for assisting with the preliminary measurements. This work was supported by research grants from VILLUM FONDEN (VIL50296), the National Science Foundation (#2339707), the Institute of Information & communications Technology Planning & Evaluation (IITP) funded by the Korea government (MSIT) (No.2019-0-01842, Artificial Intelligence Graduate School Program (GIST)), and the Culture, Sports and Tourism R&D Program through the Korea Creative Content Agency funded by the Ministry of Culture, Sports and Tourism in 2023 (RS-2023-00226263).

References

  • (1)
  • Brown et al. (2022) Eddie Brown, David R Large, Hannah Limerick, William Frier, and Gary Burnett. 2022. Augmenting automotive gesture infotainment interfaces through mid-air haptic icon design. In Ultrasound Mid-Air Haptics for Touchless Interfaces. Springer, 119–145.
  • Carter et al. (2013) Tom Carter, Sue Ann Seah, Benjamin Long, Bruce Drinkwater, and Sriram Subramanian. 2013. UltraHaptics: multi-point mid-air haptic feedback for touch surfaces. In Proceedings of the 26th annual ACM symposium on User interface software and technology. 505–514.
  • Chilles et al. (2019) Jamie Chilles, William Frier, Abdenaceur Abdouni, Marcello Giordano, and Orestis Georgiou. 2019. Laser doppler vibrometry and fem simulations of ultrasonic mid-air haptics. In 2019 IEEE World Haptics Conference (WHC). IEEE, 259–264.
  • Dalsgaard et al. (2022) Tor-Salve Dalsgaard, Joanna Bergström, Marianna Obrist, and Kasper Hornbæk. 2022. A user-derived mapping for mid-air haptic experiences. International Journal of Human-Computer Studies 168 (2022), 102920.
  • Freeman and Wilson (2021) Euan Freeman and Graham Wilson. 2021. Perception of ultrasound haptic focal point motion. In Proceedings of the 2021 International Conference on Multimodal Interaction. 697–701.
  • Frier et al. (2018) William Frier, Damien Ablart, Jamie Chilles, Benjamin Long, Marcello Giordano, Marianna Obrist, and Sriram Subramanian. 2018. Using spatiotemporal modulation to draw tactile patterns in mid-air. In Haptics: Science, Technology, and Applications: 11th International Conference, EuroHaptics 2018, Pisa, Italy, June 13-16, 2018, Proceedings, Part I 11. Springer, 270–281.
  • Hajas et al. (2020) Daniel Hajas, Dario Pittera, Antony Nasce, Orestis Georgiou, and Marianna Obrist. 2020. Mid-air haptic rendering of 2D geometric shapes with a dynamic tactile pointer. IEEE transactions on haptics 13, 4 (2020), 806–817.
  • Harrington et al. (2018) Kyle Harrington, David R Large, Gary Burnett, and Orestis Georgiou. 2018. Exploring the use of mid-air ultrasonic feedback to enhance automotive user interfaces. In Proceedings of the 10th international conference on automotive user interfaces and interactive vehicular applications. 11–20.
  • Hoshi et al. (2009) Takayuki Hoshi, Takayuki Iwamoto, and Hiroyuki Shinoda. 2009. Non-contact tactile sensation synthesized by ultrasound transducers. In World Haptics 2009-Third Joint EuroHaptics conference and Symposium on Haptic Interfaces for Virtual Environment and Teleoperator Systems. IEEE, 256–260.
  • Howard et al. (2023) Thomas Howard, Karina Driller, William Frier, Claudio Pacchierotti, Maud Marchal, and Jessica Hartcher-O’Brien. 2023. Gap detection in pairs of ultrasound mid-air vibrotactile stimuli. ACM Transactions on Applied Perceptions 20, 1 (2023), 1–17.
  • Howard et al. (2019) Thomas Howard, Gerard Gallagher, Anatole Lécuyer, Claudio Pacchierotti, and Maud Marchal. 2019. Investigating the recognition of local shapes using mid-air ultrasound haptics. In 2019 IEEE World Haptics Conference (WHC). IEEE, 503–508.
  • Howard et al. (2022) Thomas Howard, Maud Marchal, and Claudio Pacchierotti. 2022. Ultrasound mid-air tactile feedback for immersive virtual reality interaction. In Ultrasound Mid-Air Haptics for Touchless Interfaces. Springer, 147–183.
  • Hung et al. (2013) Gary MY Hung, Nigel W John, Chris Hancock, Derek A Gould, and Takayuki Hoshi. 2013. UltraPulse-simulating a human arterial pulse with focussed airborne ultrasound. In 2013 35th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC). IEEE, 2511–2514.
  • Hung et al. (2014) Gary MY Hung, Nigel W John, Chris Hancock, and Takayuki Hoshi. 2014. Using and validating airborne ultrasound as a tactile interface within medical training simulators. In Biomedical Simulation: 6th International Symposium, ISBMS 2014, Strasbourg, France, October 16-17, 2014. Proceedings 6. Springer, 30–39.
  • Hwang et al. (2017a) Inwook Hwang, Jeongil Seo, and Seungmoon Choi. 2017a. Perceptual space of superimposed dual-frequency vibrations in the hands. PloS one 12, 1 (2017), e0169570.
  • Hwang et al. (2017b) Inwook Hwang, Hyungki Son, and Jin Ryong Kim. 2017b. AirPiano: Enhancing music playing experience in virtual reality with mid-air haptic feedback. In 2017 IEEE world haptics conference (WHC). IEEE, 213–218.
  • Lim and Park (2023) Chungman Lim and Gunhyuk Park. 2023. Can a Computer Tell Differences between Vibrations?: Physiology-Based Computational Model for Perceptual Dissimilarity Prediction. In Proceedings of the 2023 CHI Conference on Human Factors in Computing Systems. 1–15.
  • Limerick et al. (2019) Hannah Limerick, Richard Hayden, David Beattie, Orestis Georgiou, and Jörg Müller. 2019. User engagement for mid-air haptic interactions with digital signage. In Proceedings of the 8th ACM international symposium on pervasive displays. 1–7.
  • Mulot et al. (2023a) Lendy Mulot, Thomas Howard, Claudio Pacchierotti, and Maud Marchal. 2023a. Improving the Perception of Mid-Air Tactile Shapes With Spatio-Temporally-Modulated Tactile Pointers. ACM Transactions on Applied Perception 20, 4 (2023), 1–16.
  • Mulot et al. (2023b) Lendy Mulot, Thomas Howard, Claudio Pacchierotti, and Maud Marchal. 2023b. Ultrasound Mid-Air Haptics for Hand Guidance in Virtual Reality. IEEE Transactions on Haptics (2023).
  • Obrist et al. (2013) Marianna Obrist, Sue Ann Seah, and Sriram Subramanian. 2013. Talking about tactile experiences. In Proceedings of the SIGCHI conference on human factors in computing systems. 1659–1668.
  • Obrist et al. (2015) Marianna Obrist, Sriram Subramanian, Elia Gatti, Benjamin Long, and Thomas Carter. 2015. Emotions mediated through mid-air haptics. In Proceedings of the 33rd annual ACM conference on human factors in computing systems. 2053–2062.
  • Park and Choi (2011) Gunhyuk Park and Seungmoon Choi. 2011. Perceptual space of amplitude-modulated vibrotactile stimuli. In 2011 IEEE world haptics conference. IEEE, 59–64.
  • Rakkolainen et al. (2020) Ismo Rakkolainen, Euan Freeman, Antti Sand, Roope Raisamo, and Stephen Brewster. 2020. A survey of mid-air ultrasound haptics and its applications. IEEE Transactions on Haptics 14, 1 (2020), 2–19.
  • Raza et al. (2019) Ahsan Raza, Waseem Hassan, Tatyana Ogay, Inwook Hwang, and Seokhee Jeon. 2019. Perceptually correct haptic rendering in mid-air using ultrasound phased array. IEEE Transactions on Industrial Electronics 67, 1 (2019), 736–745.
  • Rutten et al. (2020) Isa Rutten, William Frier, and David Geerts. 2020. Discriminating between intensities and velocities of mid-air haptic patterns. In Haptics: Science, Technology, Applications: 12th International Conference, EuroHaptics 2020, Leiden, The Netherlands, September 6–9, 2020, Proceedings 12. Springer, 78–86.
  • Rutten et al. (2019) Isa Rutten, William Frier, Lawrence Van den Bogaert, and David Geerts. 2019. Invisible touch: How identifiable are mid-air haptic shapes?. In Extended abstracts of the 2019 CHI conference on human factors in computing systems. 1–6.
  • Seifi et al. (2015) Hasti Seifi, Kailun Zhang, and Karon E MacLean. 2015. VibViz: Organizing, visualizing and navigating vibration libraries. In 2015 IEEE World Haptics Conference (WHC). IEEE, 254–259.
  • Shen et al. (2023) Zhouyang Shen, Madhan Kumar Vasudevan, Jan Kučera, Marianna Obrist, and Diego Martinez Plasencia. 2023. Multi-point STM: Effects of Drawing Speed and Number of Focal Points on Users’ Responses using Ultrasonic Mid-Air Haptics. In Proceedings of the 2023 CHI Conference on Human Factors in Computing Systems. 1–11.
  • Takahashi et al. (2019) Ryoko Takahashi, Keisuke Hasegawa, and Hiroyuki Shinoda. 2019. Tactile stimulation by repetitive lateral movement of midair ultrasound focus. IEEE transactions on haptics 13, 2 (2019), 334–342.
  • Vi et al. (2017) Chi Thanh Vi, Damien Ablart, Elia Gatti, Carlos Velasco, and Marianna Obrist. 2017. Not just seeing, but also feeling art: Mid-air haptic experiences integrated in a multisensory art exhibition. International Journal of Human-Computer Studies 108 (2017), 1–14.
  • Villa et al. (2022) Steeven Villa, Sven Mayer, Jess Hartcher-O’Brien, Albrecht Schmidt, and Tonja-Katrin Machulla. 2022. Extended mid-air ultrasound haptics for virtual reality. Proceedings of the ACM on Human-Computer Interaction 6, ISS (2022), 500–524.
  • Wilson et al. (2014) Graham Wilson, Thomas Carter, Sriram Subramanian, and Stephen A Brewster. 2014. Perception of ultrasonic haptic feedback on the hand: localisation and apparent motion. In Proceedings of the SIGCHI conference on human factors in computing systems. 1133–1142.
  • Wojna et al. (2023) Katarzyna Wojna, Orestis Georgiou, David Beattie, William Frier, Michael Wright, and Christof Lutteroth. 2023. An Exploration of Just Noticeable Differences in Mid-Air Haptics. In 2023 IEEE World Haptics Conference (WHC). IEEE, 410–416.
  • Yoo et al. (2022) Yongjae Yoo, Inwook Hwang, and Seungmoon Choi. 2022. Perceived intensity model of dual-frequency superimposed vibration: Pythagorean sum. IEEE Transactions on Haptics 15, 2 (2022), 405–415.