Prediction of Metacarpophalangeal joint angles and Classification of Hand configurations based on Ultrasound Imaging of the Forearm
Abstract
With the advancement in computing and robotics, it is necessary to develop fluent and intuitive methods for interacting with digital systems, augmented/virtual reality (AR/VR) interfaces, and physical robotic systems. Hand movement recognition is widely used to enable this interaction. Hand configuration classification and Metacarpophalangeal (MCP) joint angle detection are important for a comprehensive reconstruction of the hand motion. Surface electromyography (sEMG) and other technologies have been used for the detection of the hand motions. Ultrasound images of the forearm offer a way to visualize the internal physiology of the hand from a musculoskeletal perspective. Recent work has shown that these images can be classified using machine learning to predict various hand configurations. In this paper, we propose a Convolutional Neural Network (CNN) based deep learning pipeline for predicting the MCP joint angles. We supplement our results by using a Support Vector Classifier (SVC) to classify the ultrasound information into several predefined hand configurations based on activities of daily living (ADL). Ultrasound data from the forearm was obtained from 6 subjects who were instructed to move their hands according to predefined hand configurations relevant to ADLs. Motion capture data was acquired as the ground truth for hand movements at different speeds (0.5 Hz, 1 Hz, & 2 Hz) for the index, middle, ring, and pinky fingers. We were able to get promising SVC classification results on a subset of our collected data set. We demonstrated a correspondence between the predicted MCP joint angles and the actual MCP joint angles for the fingers, with an average root mean square error of 7.35 degrees. We implemented a low latency (6.25 - 9.1 Hz) pipeline for the prediction of both MCP joint angles and hand configuration estimation aimed at real-time control of digital devices, AR/VR interfaces, and physical robots.
Keywords: AI-Enabled Robotics; Gesture, Posture and Facial Expressions; Wearable Robotics
I Introduction
Smart and intuitive interaction with physical and non-physical worlds (AR/VR) is a topic of great interest to the human-computer interaction research community and has numerous interfacing and control applications. Upper limb motor dexterity in humans is possible with a highly advanced neuromuscular machinery developed over millions of years of evolution from the primitive primates with prehensile appendages to our current upper limbs. For humans, our upper limbs have the maximum proportion of motor and sensory innervation in the human body and we use them to manipulate and interact with the physical world. There is a growing interest in developing technologies that can facilitate a fluent and intuitive upper limb interaction and interfacing with digital, AR/VR and remote domains such as teleoperated robots [1]. Several physical sensors and vision based techniques have been utilized to obtain the MCP joint angles which can then be reconstructed in digital, physical and AR/VR environments [2]. Placing sensors on the fingers, such as bend sensors and motion sensors can limit mobility and usage of the hands to their full extent for the individuals wearing such sensors. Vision based methods for hand state and orientation detection are dependent on the light condition and line of sight. Therefore, biological signals from the forearms can be used as a better alternative to understanding hand movement without impeding the user and preventing difficult and awkward hand movements.

I-A Related Works
Various contact based and non-contact based technologies have been used to understand hand motion. These include sEMG [3], FMG [4], vision based approaches [5], resistive hand gloves [6], depth information based approaches [7] and WiFi sensing [8]. sEMG has been primarily widely used for understanding and decoding the hand movements [9]. It has been researched for its use as a control input to prosthetics and powered orthotics [10]. Various combinations of sensors with sEMG has also been explored for upper and lower limb motion prediction [11, 12, 13, 14]. Araki et al. used sEMG to measure middle finger joint angle and used it to control a robotic finger [15]. Shrirao et al. used sEMG to predict index finger joint angle while performing flexion-extension rotation of the index finger [16]. Wang et al. used sEMG to predict joint angles during different grasps [17]. Deep Learning has also been used for facilitating sEMG based human computer interaction [18]. sEMG is sensitive to muscle fatigue as a consequence of long term muscle movements, and works best for fast movements of the hand which can inhibit human-like control of digital and virtual interfaces, and prosthetics/orthotics [9, 16].
While sEMG can be used to give an estimate of the muscle activation for hand and finger movements, ultrasound imaging of the forearm can be used to visualize the muscles which can be used for hand state prediction using image processing techniques. Ultrasound imaging of the forearm, or Sonomyography, has been explored as an alternative sensing modality that can capture both muscle configuration and movement [19, 20, 21]. It has been shown to be capable of identifying different hand gestures and finger movements by analyzing the image obtained from ultrasound data with a combination of image processing and classification algorithms [19, 22]. Sonomyography has also been used to classify several grasp types and controlling robotic mechanisms [19]. Akhlagi et al. demonstrated classification of various hand gestures using Nearest Neighbor classification algorithms [23]. Wrist motion classification has also been done utilizing ultrasound imaging [24]. McIntosh et al. classified 10 hand gestures based on small ultrasound data sizes using support vector machines (SVMs) and multi layer perceptrons (MLPs) [25].
I-B Contributions
While hand gesture recognition has been implemented previously, there remains to be a need for continuous prediction of the MCP joint angles for potential usage as a direct or proportional control of systems and environments. With the advances in deep learning, newer algorithms and techniques can be used to make useful predictions from ultrasound images. We can use this to predict finer hand movements primarily the MCP joint angles while attaining hand configurations relevant to ADL. In this paper, we present our work on utilizing a CNN for predicting MCP joint angles based on ultrasound imaging. We supplement our implementation by using an SVC for classifying different hand motions. We also include results from our simultaneous state/angle prediction pipeline. This work is the foundation of our future research in the domain of human-computer and human-robot interaction that can be facilitated by ultrasound imaging.
II Methods
We used an SVC for hand state classification and a CNN for predicting finger angles based on the ultrasound images from the forearm. We also developed a low-latency demo to show the performance of our combined hand state classification and finger angle prediction pipeline.
II-A Hand Configuration Classification
SVMs are a class of classical machine learning algorithms based on supervised learning which are used for image classification and regression problems. For ultrasound forearm image classification to predict hand configurations, SVC and other classical machine learning algorithms have been explored in the past [19, 25]. A support vector classifier constructs hyperplanes in high-dimensional spaces, wherein a good separation is achieved by the hyperplane that has the largest distance to the nearest training-data point of any class.
II-B MCP joint angle estimation
MCP joint angle prediction was implemented using a CNN. CNNs are a class of deep learning algorithms popularly used for two dimensional image data regression based prediction. The CNN architecture that was used in the paper is based on VGG16 [26].
II-C Combined System Demo
The combined system demonstration pipeline uses a single script to use saved SVC and CNN models to predict both the hand configuration and MCP joint angles. First, the ultrasound image files are loaded and SVC model is loaded. The SVC model prediction decides the model to be loaded for the CNN. Once the CNN model is loaded, the CNN prediction is executed and the four finger MCP angle prediction plot is displayed. This is done repeatedly for a given number of frames which can be specified by the user as a way to demonstrating the combined pipeline shown in Fig 1.


III Experimental Implementation
Ultrasound data was captured from the right forearm of 6 subjects (1 female, 5 males; Age: 22.83 3.14 years; Height: 173.55 11.02 cm; Probe placement forearm diameter: 20.43 2.74 cm). Three dimensional positional data of specified points on their index, middle, ring and pinky fingers was acquired along with the ultrasound data, with the ultrasound probe mounted in a transverse position on the forearm. This data was acquired at three different speeds of hand movement.
The experimental setup uses a Vantage 128 Verasonics ultrasound data acquisition system (Verasonics, WA, USA) and a Vicon Nexus motion capture system (Vicon Motion Systems Ltd., UK). The detailed data acquisition and processing pipeline is shown in Fig 2.
III-A Human Subjects and IRB Approval
The study was approved by the institutional research ethics committee at the Worcester Polytechnic Institute (No. IRB-21-0452), and written informed consent was given by the subjects prior to all sessions. Table I lists the age, sex, height and forearm diameter at the ultrasound probe location for the 6 subjects enrolled in the study.
S. #1 | S. #2 | S. #3 | S. #4 | S. #5 | S. #6 | |
---|---|---|---|---|---|---|
Age (yr) | 25 | 20 | 24 | 23 | 24 | 21 |
Sex (M/F) | F | M | M | M | M | M |
Height (cm) | 157.5 | 170.2 | 172.7 | 180.3 | 167.6 | 193.0 |
Forearm Dia. (cm) | 22.9 | 18.4 | 18.4 | 19.1 | 18.4 | 25.4 |
III-B Instrumentation
Velcro straps on both sides of the custom designed 3D printed probe casing were used to strap the ultrasound probe on the forearm. The subject’s arm rests on a rest table which was further secured with another external Velcro extender to the table after verification of the ultrasound images. Figure 5 shows a subject’s right arm strapped to the rest table.
III-B1 Verasonics Ultrasound System
An L12/5 50 mm linear array ultrasound probe was used with the Verasonics Vantage 128/128 research ultrasound system. A MATLAB script was used to set sequence objects for the Verasonics data acquisition hardware to display and record ultrasound images. A trigger output signal was sent to an Arduino board from the Verasonics system. The ultrasound data frame acquisition rate was set at 25 Hz. For 56 seconds of data acquisition per session, 1400 frames were recorded. Each data frame was 636 x 256 pixels.
III-B2 Vicon Motion Capture System
A 10 Camera Vicon Vantage motion capture system with Lock+ 64-channel ADC (Analog to Digital Converter) was used to get the three dimension positional information for different motion capture markers attached to the fingers. The system is present in the Motion Capture Suite in Worcester Polytechnic Institute’s PracticePoint facility.

III-B3 Synchronization of Vicon and Verasonics systems
Because of a sample rate mismatch between the trigger from the Verasonics ultrasound system and the Vicon motion capture system, an Arduino Uno was used as a bridge to record trigger signals utilizing Arduino’s interrupt service routine and generate a pulse that could be read by the Vicon analog to digital converter.
III-B4 Audio Signal Generation
Three audio waveforms were designed for simulating three different speeds for hand movements using Audacity, an open source digital audio workstation. 0.1 second Sawtooth waves at 440 Hz and 392 Hz were used to alert the subjects to alternate between the rest and motion hand states. The frequencies of the rest/motion switching were set as 2 Hz for fast, 1 Hz for medium & 0.5 Hz for slow speeds.
III-B5 Computation Hardware and Software
NVIDIA GeForce RTX 2070 SUPER was used to train and run the angle prediction regression models. AMD Ryzen 7 2700X Eight-Core Processor was used to run and train the hand configuration classification models. The system had 31.91 GB available RAM. The code was executed in Python 3.7. TensorFlow Keras API was used to train and run deep learning models [27]. For SVC, scikit-learn library was used to run and train the machine learning based classification models [28].
III-C Experimental Testing Protocol
11 hand configurations relevant to the activities of daily living (ADLs) were chosen for this project [29]. The hand configurations are shown and described in the Fig 4. The Open (all fingers extended) was considered as the rest state for all the 11 hand configurations and the subjects alternated between the open hand and selected the hand configuration. The subject was seated and 16 motion capture markers were attached to their right hand using a double sided tape. Motion capture markers were attached on approximate ends of metacarpal and proximal phalanx for each of the index, middle, ring and pinky fingers as seen in Fig. 3 (a).
This marker configuration allowed us to calculate the MCP joint angle while ensuring free movement of the hand for the desired hand movements. The MCP joint angle was measured by taking the inverse cosine of the two vectors formed by each metacarpal-proximal phalanx pair. Ultrasound gel was applied to the subject’s forearm and the imaging surface on the ultrasound probe. The ultrasound probe was then strapped to the subject’s right forearm. Then, the subject’s forearm was fastened to a rest table. The subject then performed a trial run for the 12 hand movements. Auditory beeps were used to indicate the desired rate of hand opening and closing. There was a rest of 1 minute between each data acquisition session and 2 minutes between each speed transition. 56 seconds of data was acquired for each data acquisition session. Per subject, there were 33 data acquisition sessions, for 11 different hand movements at 3 different speeds each.


IV Data Analysis
Ultrasound data obtained was preprocessed before training the models by normalizing, log-compressing and reshaping to highlight relevant muscle features. For the 6 subjects we collected data for 11 hand configurations, each at 3 speeds, leading to a total of 198 sessions of data acquisition. Each file was was split into 30% test data and 70% training data.
IV-A Data and Model Parameters
For each subject for each hand configuration and speed, 4 seconds of the training and 2 seconds of test data were saved. This data was then combined across the speeds for each subject and hand configuration leading to 12 seconds of training and 6 seconds of test data per hand configuration. All the individual training and testing files were combined and then were shuffled and split with a test-train split of 0.3 to minimize any bias. For our SVC implementation, we used a linear kernel because of the large number of features in each frame of the ultrasound image. We shuffled the data and set a test-train split of 0.3.
The CNNs were trained for each hand configuration and speed on 56 seconds of data for every subject. We used the Adam optimizer function as the gradient descent method [30]. A learning rate of was set for the Adam optimizer. The learning rate decay was set to . We used mean absolute error as the loss, which computes the mean of absolute difference between labels and predictions. Each data file was split with a test-train split of 0.3 without shuffling. While training, the validation split was defined as 0.1. The number of epochs was set to 50.
IV-B Angle Estimation for Ground Truth
Motion capture markers were placed on the subject’s hand at the base and the head of the proximal phalanges and the metacarpals for the index, middle, ring and pinky fingers. The marker locations on a subject’s right hand can be seen in Fig 3(a). Data from the thumb was omitted and will be the subject of future research.
For the index metacarpal, the points can be defined as and , with the vector between these points, , shown in red in Fig 3(a). For the index proximal phalanx, the points can be defined as and , with the vector between these points, , shown in blue in Fig 3(a). These locations are based on the finger joints on the hand. By taking the dot product of the metacarpal and phalanges vectors extrapolated from the marker data, the MCP joint angles can be obtained. In Fig 3(b), the vectors and markers are shown for four different angles.
IV-C Quantification Metrics
For classification performance, classification accuracy percentage was used as the evaluation metric. It is defined as
where, is the number of True Positives, is the number of True Negatives and is the Total Sample Size. is an outcome where the model correctly predicts the positive class. is an outcome where the model correctly predicts the negative class. Root Mean Squared Error () values were used to quantify the deep learning regression results. Root Mean Squared Error is defined as
where, is the Total Sample Size, is the integer value ranging from 1 to the total number of samples, is the test data value for the sample , and is the value predicted by the deep learning algorithm at the sample .
V Results
Here we describe the hand configuration classification results using SVMs and MCP joint angle estimation results using CNNs. We also describe our successful low latency MCP joint angle prediction and hand configuration classification aimed at real time control of interfaces.
V-A Hand Configuration Classification
V-B Angle Estimation
MCP joint angle prediction was done using VGG16, a CNN network used for to obtain relevant information from image data. Utilizing VGG16, we were able to generate a good degree of visual correspondence between our test data and the model predictions. This is shown for four second test data in Fig 6 (b) through Fig 6 (h). RMSE values were obtained from each of the hand configurations for each speed, for every subject. For the results, the range is set from 0°-100°to encapsulate the angle data from all subjects. The overall RMSE value averaged over all the hand configurations, three different speeds and all the subjects was found to be of 7.35°.
V-B1 Viability for different hand configurations
RMSE values were calculated for the predictions averaged over all the subjects for all the slow, medium and fast speeds for each hand configuration. Fig 7 shows the plot of RMSE results for different hand configurations. The lowest average RMSE value (4.55°) was obtained C11 (Hook) and the largest average RMSE value (11.43°) was obtained for C10 (Fist).
V-B2 Repeatability over subjects
RMSE values were calculated for the predictions averaged over all the slow, medium and fast speeds, and the hand configuration for each subject. Fig 8 shows the plot of RMSE results for different subjects. The lowest average RMSE value (5.65°) was obtained from subject 4 and the largest average RMSE value (9.19°) was obtained from subject 3.
V-C Combined Prediction Pipeline
For the combined prediction pipeline SVC models trained to classify between 11 different hand configurations for a single speed and subject and 11 VGG16 models trained for prediction of MCP joint angles were used. In the combined loop, first the hand configuration state prediction is done using the saved SVC model. Then, based on the SVC prediction, the CNN model is chosen from a dictionary of saved models. For each frame of data, the SVC prediction takes 0.1 - 0.15 seconds. CNN prediction takes 0.004 - 0.007 seconds. The total time ranges from 0.11 - 0.16 seconds for angle and state prediction for one frame of ultrasound information, leading to 6.25 - 9.1 Hz of data processing capability using the combined prediction pipeline.


VI Discussion and Future Work
Our experiments show that it is possible to obtain both hand configuration classification as well as finger angle estimation using one frame of ultrasound data using different classification and regression models. While ultrasound hand configuration classification is a well researched task, we were able to show that with our experimental setup we were able to obtain 100% accuracy on ultrasound data combined for different speeds. These results were consistent with the high classification accuracy obtained by McIntosh et al. [25]. Due to computational limitations, we were not able to train classifiers which were based on ultrasound data which was more than 198 seconds in length. It would be worthwhile to check out the classifier performance for bigger data-sets and to see how deep learning based classification performs on ultrasound data in comparison with the traditional SVC. For real time control applications, it would be necessary to evaluate the classification in an on-the-fly training and evaluation paradigm, similar to the 4 state discretized classifier trained in [19]. We trained our SVC model on data acquired in a way that motion artifacts are minimised, and the subjects were supervised in a way that they would be consistent with their movements for a particular hand configuration movement set which lead to obtaining 100% classification accuracy on our combined speed data.
During the data acquisition, because of erratic hand movements and motion capture marker occlusion, there were issues with getting proper ground truth for some files. Because of these issues, we were only able to get the RMSE results for 70% of the total data files that we initially acquired. As can be seen in Fig 7, configurations C4 and C10 have the biggest RMSE values. This is because both these states (PinFlex and Fist) have significant pinky finger movement. Pinky finger movements were not consistent because it’s hard to have repeatable movements with pinky finger. For configuration C11 (Hook), the RMSE value was the minimum because the MCP angle values are smaller than other hand configurations.
For the combined prediction pipeline, most of the time is taken by SVC prediction since the current implementation is not optimized on GPU and it uses the system CPU which is much slower. We believe that exploring more GPU supported deep neural network based classification algorithms can significantly improve the rate of simultaneous prediction from the current 6.25 - 9.1 Hz. In the current paper, different deep learning models were trained for each angle estimation data. It would be worthwhile to explore deep learning algorithms which can generalise the MCP joint angle prediction for different hand configurations as well as for different subjects.
VII Conclusions
In this paper, we shared our results for MCP joint angle estimation and hand state prediction based on ultrasound data acquired from the forearm. Our CNN based model was able to get an RMSE with an average value of 7.35° over 6 subjects and 11 hand configurations. Despite our limited compute for SVC, we were able to get promising classification results. We presented our combined pipeline for angle and hand state estimation which, on our current setup can generate predictions as fast as 6.25 - 9.1 Hz per frame of ultrasound data. These results are encouraging and they show the capability of ultrasound in understanding muscle movements. We hope that this work can inspire research in the promising domain of utilizing ultrasound for predicting both continuous and discrete hand movements which can be useful for an intuitive and adaptable control of physical robots and non-physical digital and AR/VR interfaces.
References
- [1] Anna Tarasenko, Mikheil Oganesyan, Daryna Ivaskevych, Sergii Tukaiev, Dauren Toleukhanov, and Nickolai Vysokov. Artificial intelligence, brains, and beyond: Imperial college london neurotechnology symposium, 2020. Bioelectricity, 2(3):310–313, 2020.
- [2] Ming Jin Cheok, Zaid Omar, and Mohamed Hisham Jaward. A review of hand gesture and sign language recognition techniques. International Journal of Machine Learning and Cybernetics, 10(1):131–153, 2019.
- [3] Md Rezwanul Ahsan, Muhammad I Ibrahimy, Othman O Khalifa, et al. Emg signal classification for human computer interaction: a review. European Journal of Scientific Research, 33(3):480–501, 2009.
- [4] Xianta Jiang, Lukas-Karim Merhi, Zhen Gang Xiao, and Carlo Menon. Exploration of force myography and surface electromyography in hand gesture classification. Medical engineering & physics, 41:63–73, 2017.
- [5] Munir Oudah, Ali Al-Naji, and Javaan Chahl. Hand gesture recognition based on computer vision: a review of techniques. journal of Imaging, 6(8):73, 2020.
- [6] Lin Guo, Zongxing Lu, and Ligang Yao. Human-machine interaction sensing technology based on hand gesture recognition: A review. IEEE Transactions on Human-Machine Systems, 2021.
- [7] Jesus Suarez and Robin R Murphy. Hand gesture recognition with depth images: A review. In 2012 IEEE RO-MAN: the 21st IEEE international symposium on robot and human interactive communication, pages 411–417. IEEE, 2012.
- [8] Hasmath Farhana Thariq Ahmed, Hafisoh Ahmad, and CV Aravind. Device free human gesture recognition using wi-fi csi: A survey. Engineering Applications of Artificial Intelligence, 87:103281, 2020.
- [9] Kexiang Li, Jianhua Zhang, Lingfeng Wang, Minglu Zhang, Jiayi Li, and Shancheng Bao. A review of the key technologies for semg-based human-robot interaction systems. Biomedical Signal Processing and Control, 62:102074, 2020.
- [10] Daniel Graupe, Javad Salahi, and Kate H Kohn. Multifunctional prosthesis and orthosis control via microcomputer identification of temporal pattern differences in single-site myoelectric signals. Journal of Biomedical Engineering, 4(1):17–22, 1982.
- [11] Ung Hee Lee, Justin Bi, Rishi Patel, David Fouhey, and Elliott Rouse. Image transformation and cnns: A strategy for encoding human locomotor intent for autonomous wearable robots. IEEE Robotics and Automation Letters, 5(4):5440–5447, 2020.
- [12] Nadav D Kahanowich and Avishai Sintov. Robust classification of grasped objects in intuitive human-robot collaboration using a wearable force-myography device. IEEE Robotics and Automation Letters, 6(2):1192–1199, 2021.
- [13] Wentao Dong, Lin Yang, Raffaele Gravina, and Giancarlo Fortino. Soft wrist-worn multi-functional sensor array for real-time hand gesture recognition. IEEE Sensors Journal, 2021.
- [14] Kaitlin G Rabe, Mohammad Hassan Jahanandish, Kenneth Hoyt, and Nicholas P Fey. Use of sonomyographic sensing to estimate knee angular velocity during varying modes of ambulation. In 2020 42nd Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC), pages 3799–3802. IEEE, 2020.
- [15] Nozomu Araki, Kenji Inaya, Yasuo Konishi, and Kunihiko Mabuchi. An artificial finger robot motion control based on finger joint angle estimation from emg signals for a robot prosthetic hand system. In The 2012 international conference on advanced mechatronic systems, pages 109–111. IEEE, 2012.
- [16] Nikhil A Shrirao, Narender P Reddy, and Durga R Kosuri. Neural network committees for finger joint angle estimation from surface emg signals. Biomedical engineering online, 8(1):1–11, 2009.
- [17] Chao Wang, Weiyu Guo, Hang Zhang, Linlin Guo, Changcheng Huang, and Chuang Lin. semg-based continuous estimation of grasp movements by long-short term memory network. Biomedical Signal Processing and Control, 59:101774, 2020.
- [18] Dezhen Xiong, Daohui Zhang, Xingang Zhao, and Yiwen Zhao. Deep learning for emg-based human-machine interaction: A review. IEEE/CAA Journal of Automatica Sinica, 8(3):512–533, 2021.
- [19] Keshav Bimbraw, Elizabeth Fox, Gil Weinberg, and Frank L Hammond. Towards sonomyography-based real-time control of powered prosthesis grasp synergies. In 2020 42nd Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC), pages 4753–4757. IEEE, 2020.
- [20] Yong-Ping Zheng, MMF Chan, Jun Shi, Xin Chen, and Qing-Hua Huang. Sonomyography: Monitoring morphological changes of forearm muscles in actions with the feasibility for the control of powered prosthesis. Medical engineering & physics, 28(5):405–415, 2006.
- [21] Xin Chen, Yong-Ping Zheng, Jing-Yi Guo, and Jun Shi. Sonomyography (smg) control for powered prosthetic hand: a study with normal subjects. Ultrasound in medicine & biology, 36(7):1076–1088, 2010.
- [22] Jun Shi, Qian Chang, Yong-Ping Zheng, et al. Feasibility of controlling prosthetic hand using sonomyography signal in real time: preliminary study. 2010.
- [23] Nima Akhlaghi, Clayton A Baker, Mohamed Lahlou, Hozaifah Zafar, Karthik G Murthy, Huzefa S Rangwala, Jana Kosecka, Wilsaan M Joiner, Joseph J Pancrazio, and Siddhartha Sikdar. Real-time classification of hand motions using ultrasound imaging of forearm muscles. IEEE Transactions on Biomedical Engineering, 63(8):1687–1698, 2015.
- [24] Xingchen Yang, Jipeng Yan, Yinfeng Fang, Dalin Zhou, and Honghai Liu. Simultaneous prediction of wrist/hand motion via wearable ultrasound sensing. IEEE Transactions on Neural Systems and Rehabilitation Engineering, 28(4):970–977, 2020.
- [25] Jess McIntosh, Asier Marzo, Mike Fraser, and Carol Phillips. Echoflex: Hand gesture recognition using ultrasound imaging. In Proceedings of the 2017 CHI Conference on Human Factors in Computing Systems, pages 1923–1934, 2017.
- [26] Karen Simonyan and Andrew Zisserman. Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556, 2014.
- [27] Aurélien Géron. Hands-on machine learning with Scikit-Learn, Keras, and TensorFlow: Concepts, tools, and techniques to build intelligent systems. O’Reilly Media, 2019.
- [28] Fabian Pedregosa, Gaël Varoquaux, Alexandre Gramfort, Vincent Michel, Bertrand Thirion, Olivier Grisel, Mathieu Blondel, Peter Prettenhofer, Ron Weiss, Vincent Dubourg, et al. Scikit-learn: Machine learning in python. the Journal of machine Learning research, 12:2825–2830, 2011.
- [29] Aaron M Dollar. Classifying human hand use and the activities of daily living. In The Human Hand as an Inspiration for Robot Hand Development, pages 201–216. Springer, 2014.
- [30] Diederik P Kingma and Jimmy Ba. Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980, 2014.