Enable Natural Tactile Interaction for Robot Dog based on Large-format Distributed Flexible Pressure Sensors
Abstract
Touch is an important channel for human-robot interaction, while it is challenging for robots to recognize human touch accurately and make appropriate responses. In this paper, we design and implement a set of large-format distributed flexible pressure sensors on a robot dog to enable natural human-robot tactile interaction. Through a heuristic study, we sorted out 81 tactile gestures commonly used when humans interact with real dogs and 44 dog reactions. A gesture classification algorithm based on ResNet is proposed to recognize these 81 human gestures, and the classification accuracy reaches 98.7%. In addition, an action prediction algorithm based on Transformer is proposed to predict dog actions from human gestures, reaching a 1-gram BLEU score of 0.87. Finally, we compare the tactile interaction with the voice interaction during a freedom human-robot-dog interactive playing study. The results show that tactile interaction plays a more significant role in alleviating user anxiety, stimulating user excitement and improving the acceptability of robot dogs.
I INTRODUCTION
Touch is one of the most primitive sensory channels of organisms. Harlow’s rhesus monkey experiment [1] shows that comfortable physical contact plays an important role in natural emotion development. Tactile interaction is also of great significance for the relationship between humans and pet robots. Studies have found that affective connection with pet dogs can reduce humans’ loneliness [2], and tactile contact with them can even benefit humans’ health [3]. Since the development of pet robots, they have been able to show some animal-like behaviors when interacting with humans, including affective expressions [4]. In particular, tactile interaction will affect human emotions, which will be reflected in their different gestures when touching robot dogs [5].
However, compared with the popular visual and auditory technologies, the tactile-based human-robot interaction technology is relatively immature. Previous studies in the field of robot tactile sensing mainly focus on local perception, such as mechanical arm [6, 7], gripper [8, 9], local robot skin [10], etc.; and the whole-body tactile sensing systems are primarily based on discrete [11] or low-density [12, 13] sensor layouts. Up to now, it has been challenging and seldom implemented to enable large-format and high-density tactile sensing for robots. Moreover, there are few studies on the human-robot tactile interaction pipeline, i.e., from human touch to robot feedback.
To fill these gaps, we design and implement a real-time interactive system based on large-format distributed flexible pressure sensors in this paper. By sticking a layer of piezoresistive film sensors on the skin of a robot dog, the sensors cover the robot’s skin in a large area and a special shape, including the head, back, hips, legs and other parts.
Based on existing work collecting the gesture set when humans interact with tactile creatures [5, 13, 14], we carried out a heuristic study, including questionnaire distribution and video collection, to examine humans’ gesture preference in tactile interaction with dogs and dogs’ feedback actions. Through questionnaire screening and video analysis, we sorted out 13 human gestures and 11 dog body parts, and combined them to obtain 81 tactile gestures. In addition, we summarized 44 dog actions from the videos.
We use the gesture recognition algorithm based on ResNet [15] to classify the above 81 human gestures, and the accuracy reaches 98.7%. These gestures will trigger different dog actions, which is realized by Transformer [16], reaching a 1-gram BLEU [17] score of 0.87.

Finally, we conducted a compared experiment to verify the superiority of our tactile interaction over the voice interaction on human emotion regulation. The results show that compared with the voice interaction, the participants in the tactile group have more significant emotional progress after the experiment and are willing to spend more time playing with the robot dog.
Therefore, the contributions of this paper are as follows:
-
•
A real-time interactive system based on large-format distributed flexible pressure sensors, enabling human-robot-dog tactile interaction from human touch to dog reactions.
-
•
Eighty-one human tactile gestures commonly used when interacting with dogs and 44 dog feedback actions are summarized through a heuristic study.
-
•
A novel algorithm architecture is proposed, with a human gesture classification algorithm based on ResNet and a dog action prediction algorithm based on Transformer, which achieves a classification accuracy of 98.7% and a 1-gram BLEU score of 0.87, respectively.
-
•
A compared experiment that reveals human-robot-dog tactile interaction is significantly more effective in promoting users’ positive emotions than voice interaction.
II Hardware
The hardware system consists of the robot dog, the piezoresistive film sensors and the sensor reading boards. They are described in detail below.
II-A Piezoresistive Film Sensor
The piezoresistive film sensors we use are off-shelf provided by Roxifsr111https://world.taobao.com/item/609970762251.htm. It is made by transferring the force-sensitive materials, silver paste and other materials to the flexible film substrate through the precision printing process, and then drying and curing. When pressure is applied to the sensor, the resistance decreases as the pressure increases.
As shown in Fig.1, a total of 13 pressure sensors with three sizes are applied to the robot dog: i) six 32 32 sensors are attached to the body; ii) three 13 16 sensors attached to the head, the cheeks and the chin; iii) four single-point sensors attached to the forelimbs.
The 32 32 sensor arrays provide 1024 sensing points within 16 16 cm2 area, with a maximum measurement range of 2. The 13 16 sensor arrays are cut out from the 32 32 ones, all of which share the same parameters. The single-point pressure sensors are located at the front and back of each forelimb, with a maximum measurement range of 25. Besides, the top and bottom of all sensors are wrapped with aluminum foil and connected with copper foil tape, forming a shield to reduce noise.
II-B Sensor Reading Board
There are eight sensor reading boards installed in the robot dog, six for the body sensors, one for the cheeks and chin sensors, and one for the head and forelimbs sensors. The sensor reading board adopts STM32F072R8T6 as MCU (Micro Control Unit), 8x CD4051R as multiplexer to select rows and columns, and LM321M5 as voltage amplifier. All sensor reading boards are wired to the PC and powered by external power through an 8-in-1 USB hub. They collect data at a frame rate of 10, and the serial ports realize data synchronization through broadcast communication.
II-C Robot Dog
The robot dog we use is the CyberDog provided by MI222https://www.mi.com/cyberdog. GRPC communication is adopted between the PC and CyberDog, which is guaranteed to be in a LAN through a wired connection. Tactile input is calculated as action instruction on PC, and transmitted to CyberDog for execution. The action arrangement of CyberDog is based on ROS2, and different designs are realized by modifying the action parameters of .toml files. Twelve brushless motors in the robot dog are controlled by calling the .toml file, changing the postures and realizing the action control of the robot dog. In addition, for the beautiful appearance and comfortable touch, we customized a fur cover for the robot dog.
III Heuristic Study



In this paper, we define tactile gestures as the combination of human gestures and interaction parts on the dog, such as stroking the head. These tactile gestures are dynamic and usually contain human feelings or intentions when executed. In order to explore the gesture preference of dog owners for tactile interaction with their pet dogs in real life and dogs’ different reactions to different gestures, we designed a questionnaire and collected tactile interaction videos between dog owners and their dogs for research.
III-A Participants
This experiment is aimed at people who have experience keeping dogs (who have had dogs or are keeping dogs). According to the daily interaction between these people and their pet dogs, we try to find the tactile interaction pattern between humans and robot dogs. In addition to collecting questionnaires, we contacted the users who left their contact information and expressed their willingness to participate in further experiments. They were asked to submit one or more videos of touching dogs in their daily lives, with a total duration of about 10 minutes. Finally, 109 questionnaires passed the attention test and were retained. Among them, 26 users submitted their videos, which added up to 208 minutes.
III-B Task and Procedure
According to the collected 109 questionnaires, we sorted out users’ preferred gestures and obtained a group of tactile gestures. In addition, by analyzing the 26 interaction videos, we supplemented some common gestures and extracted dog feedback actions under corresponding triggers.
III-B1 Human Tactile Gestures
Referring to the dictionary of 30 human gestures of touching tactile creatures compiled by Yohanan and MacLean [5], we summarized the tactile gesture set of human-robot-dog interaction in this paper through the following screening scheme.
-
•
First of all, we screened out four of the 30 gestures that could not be done to the robot dog because of its heavy weight: Hold, Lift, Swing and Toss.
-
•
In the questionnaire, we invited users to rate these 26 tactile gestures using the five-point Likert scale. The scores from 1 to 5 represent the increasing preference. After summary and analysis, we selected 15 gestures that all users scored more than 3 points on average.
-
•
Gestures that can not be detected by the piezoresistive sensors and are potentially harmful to the posture balance of the robot dog were also deleted, that is, Pick, Pinch, Finger Idly, Tremble and Rock.
-
•
Handhold and Handshake were added to the gesture set because they frequently appeared in videos.
-
•
The gesture “none” was added to describe the situation where no tactile interaction occurs.
Finally, a total of 13 gestures (see Fig.2) were included in the tactile gesture set. By observing the videos, we found that the dog body parts touched by users could be roughly divided into five categories: head, back, abdomen, hips and forelimbs. However, the interaction on the abdomen usually occurs when the dog lies on his back, and we have not arranged any related actions (i.e., the actions marked with gray in Fig. 4) at present because of the raised back. Therefore, for the current implementation, abdominal interaction was considered to be excluded, and the interaction parts were divided into 11 (see Fig. 3) after detailed classification. Finally, 81 human gestures towards the robot dog were compiled by combining 13 tactile gestures with 11 interaction parts.
III-B2 Dog Feedback Actions
We collected the dog action set by observing and analyzing the real reactions of dogs to human touch in the videos. The followings are our processes and rules for screening and summarizing dog actions.
-
•
Sort out the common actions repeated in different dogs. According to the changes in the position, direction and stretching degree of specific body parts of the dogs, we summarized actions like standing, sitting and lying down, etc. Besides, we concluded actions such as sticking and rubbing according to the touch strength and angle, and whether there was relative sliding.
-
•
Actions that have only appeared a few times but are interesting were also kept in our list, such as crouching slightly and hindlimbs jumping.
-
•
Screen out actions that the robot dog cannot perform or are dangerous, such as twisting the neck or pouncing on someone.
-
•
Some slight actions, such as limb shaking, were not considered.
-
•
Eliminate interactions triggered by senses other than touch, such as hearing or vision.
-
•
Screen out interactions caused by other items, not the human body.
-
•
According to the body parts and the action fineness, dog actions were divided into four categories: whole body actions, forelimbs actions, hindlimbs actions and head actions.
Finally, we came to the conclusion that dog feedback actions to human tactile gestures could be divided into four categories, with a total of 44 actions (see Fig.4).
III-C Discussion
The interaction between humans and dogs reflected in the videos is quite rich. Human gestures and dog actions usually show a reasonable correspondence, but they are not completely consistent. For example, when a person strokes a dog’s head, the dog may bow its head or rub its head against the person’s hand. These reasonable but diverse reactions are a great pleasure for humans to play with their pet dogs, which provides us with a powerful inspiration to realize the tactile feedback mechanism.
In short, based on the questionnaires and videos collected, we summarized 13 human gestures and 11 interaction parts on the dog, making a total of 81 tactile gestures. With regard to dog feedback actions to human tactile gestures, we analyzed their performance in the videos and finally compiled an action set containing 44 common actions.
IV Algorithm Architecture

The overall algorithm architecture from human tactile gestures to dog feedback actions is shown in Fig. 5, which consists of two main steps: human tactile gesture classification and dog feedback action prediction. Tactile gesture recognition is implemented using ResNet [15] architecture, with tactile image sequence as input and tactile gesture classification result as output. Dog action prediction is implemented using Transformer [16] architecture, which is actually regarded as a classic translation task. The prediction algorithm, Transformer, is used to “translate” human gesture “language” into dog action “language”, taking human gesture “word (s)” (sequence) as input and dog action “word (s)” (sequence) as output.
IV-A Tactile Gesture Recognition
IV-A1 Dataset
By combining 13 human gestures with 11 dog body parts, we get 81 categories of tactile gestures. The gesture dataset includes 14,580 image sequences, including 180 in each category. For the dataset, 90 is kept for training while the remaining is for testing. In the data collection process, each gesture was repeated 20 times by each person, with 9 people participating, i.e., 81 categories 20 times 9 people = 14,580 image sequences. Each image sequence includes 20 frames of tactile images, that is, a total of 291,600 tactile images. The resolution of each tactile image is 64 128, composed of eight 32 32 sensor data.
IV-A2 Classification Algorithm
We formulate our gesture recognition task as a multiclass (i.e., 81 classes) classification problem. Accordingly, we build a variant of the ResNet34 deep residual CNN by adjusting its input layer to match our input image size of H W C (H 64, W 128, C 20) and its output layer to match our output dimension of 81. Then, we can train our ResNet34 model in a supervised manner using our dataset. We implement this model in Python using the pytorch framework.
IV-A3 Results
Our gesture classification model achieves 98.7% accuracy on the 81-class classification task on the test set.
IV-B Dog Action Prediction
IV-B1 Vocabulary Construction
We formulate our action prediction task as a translation problem. Therefore, we have constructed two vocabularies, one for human gesture names and the other for dog action names. The former contains 81 human gestures, while the latter contains 40 dog actions.
IV-B2 Dataset
We collected a total of 208-minute videos from 26 participants. These videos showed people’s tactile gestures when they interacted with dogs and dogs’ feedback actions. We recorded the human gestures and the dog reactions respectively, and constructed a dataset of 1212 human gestures-dog actions interaction sequences. Five-sixths of the dataset is kept for training, while the remaining is for testing.
IV-B3 Prediction Algorithm
We adopt the Transformer model, which relies on the multi-head self-attention mechanism to realize the translation task of sequence to sequence. The input of the network is a sequence of human gesture words with a maximum length of 20 (including a beginning word and an ending word), representing the current and the previous human gestures to consider the context. The output of the network is the corresponding sequence of dog action words. We take the last word of the output sequence as the prediction result of the dog action.
IV-B4 Results
BLEU [17] is used to evaluate the performance of the action prediction algorithm. On the translation task from human gestures to dog actions, our prediction model achieved a 1-gram BLEU score of 0.87.
V Compared Experiment
In order to evaluate the impact of tactile interaction with the robot dog on humans’ emotions and attitudes towards robot dogs, we carried out an experiment to compare our tactile interaction scheme with the voice scheme.
V-A Participants
Sixteen participants (Female 7, avg. age 26.2 years) were invited to participate in our experiment and randomly divided into two groups, each group with 8 participants:
1) Group A (experimental group): Participants interact with the robot dog by touching it;
2) Group B (control group): Participants interact with the robot dog using voice instruction, e.g., “sit down”. The voice interaction is based on the Wizard of Oz method [18];
Our research was IRB approved by the local institution of Tsinghua University. All participants signed the informed consent form before the experiment, and they were paid accordingly after the experiment.
V-B Scales Design
We designed three scales to evaluate the participants’ emotions and attitudes toward robot dogs before and after the test.
The first one is an attitude scale of humans towards robot dogs, merged with the questions which are suitable for robot dogs of Companion Animal Bonding Scale [19], Pet Attitude Scale [20, 21] and Animal Attitude Scale [22]. It is scored by the five-point Likert scale, and the participants’ acceptance of robot dogs is positively correlated with the average score of all questions.
In order to evaluate the changes in participants’ depression and anxiety, we design the second scale, consisting of the Self-Rating Depression Scale [23] and the Self-Rating Anxiety Scale [24]. The title of this scale is visible to participants as the Emotion Self-rating Scale, rated by the four-point Likert scale. The higher the average score of all questions, the stronger the participants’ depression and anxiety.
The third scale comes from Russell’s theory, which holds that emotion comprises two bipolar and orthogonal dimensions [25, 26]. One dimension, valence, is described by the continuum from unpleasant to pleasant, and the other dimension, arousal, ranges from deactivated to activated. These two dimensions are represented by the horizontal axis and vertical axis respectively, and the positive direction of the coordinate axis indicates more positive valence and higher arousal level, with a maximum coordinate value of 3.
Finally, the four indicators indicated by these three scales: humans’ attitude towards robot dogs, humans’ depression and anxiety intensity, valence level and arousal level are used to evaluate the changes in participants’ pre-and post-test status.
V-C Procedure
The experiment consists of three parts: pre-test, formal test and post-test. In the pre-and post-test, participants filled out the above three scales, respectively. In the formal test, participants were allowed to interact freely with the robot dog. The maximum interaction time was set to 20 minutes, and the interaction between participants and the robot dog would not be interrupted within 20 minutes until they took the initiative to stop. The two groups of experiments were conducted separately, and the participants in each group participated in the experiments in turn without communication.
V-D Results
Before significance tests, the experimental data are tested for normality. The data conforming to the normal distribution is tested by T-test; otherwise, the Mann-Whitney U-Test. The null hypotheses include i) there is no difference in user indicators in the pre-and post-test; ii) there is no difference in user indicators between group A and group B.
V-D1 Humans’ Attitude towards Robot Dogs
It can be seen from Table. I that participants’ attitudes towards the robot dog have changed significantly in the pre-and post-test of the experimental group. However, the attitudes of participants in the control group hardly changed. From the significance test results between groups, it can also be seen that the attitudes of the two groups of participants are roughly the same in the pre-test, but there are significant differences in the post-test. As P4 in the experimental group said, “I thought it was a machine at first sight, but after playing for a while, I thought he was like a pet dog.” However, P7 of the control group indicated that he felt he was playing a game instead of playing with a pet.
Group A | Group B | Value | P | |
pre-test | 3.410.29 | 3.270.31 | t=0.904 | P=0.191 |
post-test | 3.880.32 | 2.990.53 | t=4.030 | P<0.001*** |
Value | t=-3.067 | t=1.292 | ||
P | P=0.004** | P=0.109 |
V-D2 Humans’ Depression and Anxiety Intensity
The results in Table II show that the depression and anxiety of the participants in the experimental group were slightly relieved after interacting in touch with the robot dog, but no change occurred in the control group. The slight change may be due to the continuity of depression and anxiety that is difficult to change significantly in a short time. Nevertheless, there were significant differences in participants’ emotions between the experimental group and the control group in the post-test while without any difference in the pre-test; that is, participants’ depression and anxiety decreased even more after tactile interaction with the robot dog. Many participants in the tactile group mentioned that they felt healed.
Group A | Group B | Value | P | |
pre-test | 1.900.46 | 1.910.36 | t=-0.030 | P=0.488 |
post-test | 1.610.22 | 1.880.31 | t=-1.937 | P=0.037* |
Value | t=1.604 | t=0.184 | ||
P | P=0.065 | P=0.428 |
V-D3 Valence Level
Tactile interaction with the robot dog can significantly make participants feel pleasant, and voice interaction can also play a slight role (see Table III). Voice interaction with the robot dog gave participants the experience of playing games, which made them feel good. However, the intimacy and surprise accompanying tactile interaction could quickly make participants’ emotions reach a higher point. P5 of the experiment group gave a “cool” exclamation all the time during the experiment.
Group A | Group B | Value | P | |
pre-test | 0.600.93 | 1.130.66 | t=-1.306 | P=0.106 |
post-test | 1.730.40 | 1.630.81 | U=30.000 | P=0.878 |
Value | U=58.500 | t=-1.358 | ||
P | P=0.003** | P=0.098 |
V-D4 Arousal Level
After tactile interaction with the robot dog, the arousal level of the participants was significantly activated. However, voice interaction has no significant impact on participants’ arousal level (see Table IV). In addition, there was an interesting finding that the arousal level of participants who gave positive comments was not always improved. Some participants mentioned that the intimate and comfortable tactile interaction makes them feel at ease and become more relaxed, accompanied by a decrease in the arousal level.
Group A | Group B | Value | P | |
pre-test | -0.161.24 | 0.040.67 | t=-0.401 | P=0.347 |
post-test | 1.051.70 | 0.461.31 | U=22.000 | P=0.328 |
Value | U=51.000 | t=-0.815 | ||
P | P=0.050* | P=0.214 |
V-E Discussion
Among the above four indexes, the tactile interaction scheme has achieved significantly better results than the voice scheme. In addition, we recorded the experimental time of the participants. The average interaction time of the experimental group was 12.5 minutes, while that of the control group was 5.5 minutes. Most participants in the control group would stop the experiment after executing all the voice instructions once. On the contrary, the action feedback from natural tactile interaction could keep participants fresh. In a word, the results of the compared experiment strongly prove that the immersion experience brought by touch is of great significance to natural human-robot interaction.
VI CONCLUSION and Future Work
VI-A Natural Tactile Interaction for Robot Dog
In this paper, we implement a human-robot-dog tactile interaction system based on large-format distributed flexible pressure sensors. Through a heuristic research, we sort out 81 human tactile gestures and 44 dog feedback actions. A novel two-step algorithm architecture is proposed: a gesture classification algorithm based on ResNet to recognize these 81 human gestures with a classification accuracy of 98.7%, and an action prediction algorithm based on Transformer to predict dog actions reaching a 1-gram BLEU score of 0.87. Finally, through a compared experiment, it is proved that compared with voice interaction, natural tactile interaction plays a more significant role in relieving user anxiety, stimulating user excitement and improving the acceptability of robot dogs. Supplementary materials on gesture definitions and custom scales are available at: https://github.com/AIR-DISCOVER/Human-Robot-Dog-Tactile-Interaction.
VI-B Future Work
VI-B1 Richer Dog Actions and Whole-Body Tactile Sensing
The robot dog’s back was padded to get more space to accommodate the sensor reading boards and wires embedded in its body. As a result, it could not perform back-pressing actions such as tumbling, and no sensor was attached to the abdomen. In addition, the actions involving a flexible neck or tail were not considered due to the dog’s rigid shape. In the future, we would explore a better hardware assembly scheme and realize richer dog actions as well as the whole-body tactile sensing system on the robot dog.
VI-B2 Multi-modal Interaction
Multi-modal interaction scheme is of great significance in improving the naturalness and efficiency of human-robot interaction. Some participants mentioned it would be great if the robot dog could understand what they said while understanding tactile gestures. Therefore, we would expand voice and vision channels to further improve the system functionality and flexibility.
VI-B3 End-to-end Model Architecture
In the current implementation, a two-step model architecture, that is, gesture recognition and action prediction, is used for the human-robot-dog tactile interaction pipeline to make it more interpretable. An end-to-end model may be a meaningful and challenging attempt in further work.
References
- [1] H. F. Harlow, “The nature of love.” American psychologist, vol. 13, no. 12, p. 673, 1958.
- [2] L. Rew, “Friends and pets as companions: Strategies for coping with loneliness among homeless youth,” Journal of child and adolescent psychiatric nursing, vol. 13, no. 3, pp. 125–132, 2000.
- [3] J. K. Vormbrock and J. M. Grossberg, “Cardiovascular effects of human-pet dog interactions,” Journal of behavioral medicine, vol. 11, no. 5, pp. 509–517, 1988.
- [4] J. Lee and G. Lee, “Designing a non-contact wearable tactile display using airflows,” in Proceedings of the 29th Annual Symposium on User Interface Software and Technology, 2016, pp. 183–194.
- [5] S. Yohanan and K. E. MacLean, “The role of affective touch in human-robot interaction: Human intent and expectations in touching the haptic creature,” International Journal of Social Robotics, vol. 4, no. 2, pp. 163–180, 2012.
- [6] L. Massari, G. Fransvea, J. D’Abbraccio, M. Filosa, G. Terruso, A. Aliperta, G. D’Alesio, M. Zaltieri, E. Schena, E. Palermo, et al., “Functional mimicry of ruffini receptors with fibre bragg gratings and deep neural networks enables a bio-inspired large-area tactile-sensitive skin,” Nature Machine Intelligence, vol. 4, no. 5, pp. 425–435, 2022.
- [7] M. S. Sarwar and K. Yamane, “Large-area conformable sensor for proximity, light touch, and pressure-based gesture recognition,” in 2021 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS). IEEE, 2021, pp. 2700–2707.
- [8] L. Zlokapa, Y. Luo, J. Xu, M. Foshey, K. Wu, P. Agrawal, and W. Matusik, “An integrated design pipeline for tactile sensing robotic manipulators,” arXiv preprint arXiv:2204.07149, 2022.
- [9] S. Li, X. Yin, C. Xia, L. Ye, X. Wang, and B. Liang, “Tata: A universal jamming gripper with high-quality tactile perception and its application to underwater manipulation,” in 2022 International Conference on Robotics and Automation (ICRA). IEEE, 2022, pp. 6151–6157.
- [10] S. Eom, P. Abbaraju, Y. Xu, B. R. Nair, and R. M. Voyles, “Embedded neuromorphic architecture for form+ function 4-d printing of robotic materials: Emulation of optimized neurons,” in 2021 IEEE International Conference on Robotics and Automation (ICRA). IEEE, 2021, pp. 12 185–12 191.
- [11] T. Shibata, T. Mitsui, K. Wada, A. Touda, T. Kumasaka, K. Tagami, and K. Tanie, “Mental commit robot and its application to therapy of children,” in 2001 IEEE/ASME International Conference on Advanced Intelligent Mechatronics. Proceedings (Cat. No. 01TH8556), vol. 2. IEEE, 2001, pp. 1053–1058.
- [12] J. Chang, K. MacLean, and S. Yohanan, “Gesture recognition in the haptic creature,” in International Conference on Human Haptic Sensing and Touch Enabled Computer Applications. Springer, 2010, pp. 385–391.
- [13] A. Flagg and K. MacLean, “Affective touch gesture recognition for a furry zoomorphic machine,” in Proceedings of the 7th International Conference on Tangible, Embedded and Embodied Interaction, 2013, pp. 25–32.
- [14] P. Wang, J. Liu, F. Hou, D. Chen, Z. Xia, and S. Guo, “Organization and understanding of a tactile information dataset tacact for physical human-robot interaction,” in 2021 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS). IEEE, 2021, pp. 7328–7333.
- [15] K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning for image recognition,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2016, pp. 770–778.
- [16] A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. N. Gomez, Ł. Kaiser, and I. Polosukhin, “Attention is all you need,” Advances in neural information processing systems, vol. 30, 2017.
- [17] K. Papineni, S. Roukos, T. Ward, and W.-J. Zhu, “Bleu: a method for automatic evaluation of machine translation,” in Proceedings of the 40th annual meeting of the Association for Computational Linguistics, 2002, pp. 311–318.
- [18] P. Green and L. Wei-Haas, “The rapid development of user interfaces: Experience with the wizard of oz method,” in Proceedings of the Human Factors Society Annual Meeting, vol. 29, no. 5. SAGE Publications Sage CA: Los Angeles, CA, 1985, pp. 470–474.
- [19] R. H. Poresky, C. Hendrix, J. E. Mosier, and M. L. Samuelson, “The companion animal bonding scale: Internal reliability and construct validity,” Psychological Reports, vol. 60, no. 3, pp. 743–746, 1987.
- [20] D. I. Templer, C. A. Salter, S. Dickey, R. Baldwin, and D. M. Veleber, “The construction of a pet attitude scale,” The Psychological Record, vol. 31, no. 3, pp. 343–348, 1981.
- [21] M. Canfield, D. Templer, K. Tangan, K. Munsell, and H. Arikawa, “Modification of the pet attitude scale,” Society and Animals, vol. 12, no. 2, 2004.
- [22] H. Herzog, S. Grayson, and D. McCord, “Brief measures of the animal attitude scale,” Anthrozoös, vol. 28, no. 1, pp. 145–152, 2015.
- [23] W. W. Zung, “A self-rating depression scale,” Archives of general psychiatry, vol. 12, no. 1, pp. 63–70, 1965.
- [24] William W.K. Zung, “A rating instrument for anxiety disorders,” Psychosomatics, vol. 12, no. 6, pp. 371–379, 1971. [Online]. Available: https://www.sciencedirect.com/science/article/pii/S0033318271714790
- [25] J. A. Russell, “A circumplex model of affect.” Journal of personality and social psychology, vol. 39, no. 6, p. 1161, 1980.
- [26] J. A. Russell, A. Weiss, and G. A. Mendelsohn, “Affect grid: a single-item scale of pleasure and arousal.” Journal of personality and social psychology, vol. 57, no. 3, p. 493, 1989.