Linguistically Motivated Sign Language Segmentation
Abstract
Sign language segmentation is a crucial task in sign language processing systems. It enables downstream tasks such as sign recognition, transcription, and machine translation. In this work, we consider two kinds of segmentation: segmentation into individual signs and segmentation into phrases, larger units comprising several signs. We propose a novel approach to jointly model these two tasks.
Our method is motivated by linguistic cues observed in sign language corpora. We replace the predominant IO tagging scheme with BIO tagging to account for continuous signing. Given that prosody plays a significant role in phrase boundaries, we explore the use of optical flow features. We also provide an extensive analysis of hand shapes and 3D hand normalization.
We find that introducing BIO tagging is necessary to model sign boundaries. Explicitly encoding prosody by optical flow improves segmentation in shallow models, but its contribution is negligible in deeper models. Careful tuning of the decoding algorithm atop the models further improves the segmentation quality.
We demonstrate that our final models generalize to out-of-domain video content in a different signed language, even under a zero-shot setting. We observe that including optical flow and 3D hand normalization enhances the robustness of the model in this context.
1 Introduction
Signed languages are natural languages used by deaf and hard-of-hearing individuals to communicate through a combination of manual and non-manual elements (Sandler and Lillo-Martin, 2006). Like spoken languages, signed languages have their own distinctive grammar, and vocabulary, that have evolved through natural processes of language development (Sandler, 2010).














Sign language transcription and translation systems rely on the accurate temporal segmentation of sign language videos into meaningful units such as signs (Santemiz et al., 2009; Renz et al., 2021a) or signing sequences corresponding to subtitle units111Subtitles may not always correspond directly to sentences. They frequently split within a sentence and could be temporally offset from the corresponding signing segments. (Bull et al., 2020b). However, sign language segmentation remains a challenging task due to the difficulties in defining meaningful units in signed languages (De Sisto et al., 2021). Our approach is the first to consider two kinds of units in one model. We simultaneously segment single signs and phrases (larger units) in a unified framework.
Previous work typically approached segmentation as a binary classification task (including segmentation tasks in audio signal processing and computer vision), where each frame/pixel is predicted to be either part of a segment or not. However, this approach neglects the intricate nuances of continuous signing, where segment boundaries are not strictly binary and often blend in reality. One sign or phrase can immediately follow another, transitioning smoothly, without a frame between them being distinctly outside (Figure 1 and §3.1).
We propose incorporating linguistically motivated cues to address these challenges and improve sign language segmentation. To cope with continuous signing, we adopt a BIO-tagging approach (Ramshaw and Marcus, 1995), where in addition to predicting a frame to be in or out of a segment, we also classify the beginning of the segment as shown in Figure 2. Since phrase segmentation is primarily marked with prosodic cues (i.e., pauses, extended sign duration, facial expressions) (Sandler, 2010; Ormel and Crasborn, 2012), we explore using optical flow to explicitly model them (§3.2). Since signs employ a limited number of hand shapes, we additionally perform 3D hand normalization (§3.3).
Evaluating on the Public DGS Corpus (Prillwitz et al., 2008; Hanke et al., 2020) (DGS stands for German Sign Language), we report enhancements in model performance following specific modifications. We compare our final models after hyperparameter optimization, including parameters for the decoding algorithm, and find that our best architecture using only the poses is comparable to the one that uses optical flow and hand normalization.
Reassuringly, we find that our model generalizes when evaluated on additional data from different signed languages in a zero-shot approach. We obtain segmentation scores that are competitive with previous work and observe that incorporating optical flow and hand normalization makes the model more robust for out-of-domain data.
Lastly, we conduct an extensive analysis of pose-based hand manipulations for signed languages (Appendix C). Despite not improving our segmentation model due to noise from current 3D pose estimation models, we emphasize its potential value for future work involving skeletal hand poses. Based on this analysis, we propose several measurable directions for improving 3D pose estimation.
Our code and models are available at https://github.com/sign-language-processing/transcription.
2 Related Work
2.1 Sign Language Detection
Sign language detection (Borg and Camilleri, 2019; Moryossef et al., 2020; Pal et al., 2023) is the task of determining whether signing activity is present in a given video frame. A similar task in spoken languages is voice activity detection (VAD) (Sohn et al., 1999; Ramırez et al., 2004), the detection of when human voice is used in an audio signal. As VAD methods often rely on speech-specific representations such as spectrograms, they are not necessarily applicable to videos.
Borg and Camilleri (2019) introduced the classification of frames taken from YouTube videos as either signing or not signing. They took a spatial and temporal approach based on VGG-16 (Simonyan and Zisserman, 2015) CNN to encode each frame and used a Gated Recurrent Unit (GRU) (Cho et al., 2014) to encode the sequence of frames in a window of 20 frames at 5fps. In addition to the raw frame, they either encoded optical-flow history, aggregated motion history, or frame difference.
Moryossef et al. (2020) improved upon their method by performing sign language detection in real time. They identified that sign language use involves movement of the body and, as such, designed a model that works on top of estimated human poses rather than directly on the video signal. They calculated the optical flow norm of every joint detected on the body and applied a shallow yet effective contextualized model to predict for every frame whether the person is signing or not.
While these recent detection models achieve high performance, we need well-annotated data including interference and non-signing distractions for proper real-world evaluation. Pal et al. (2023) conducted a detailed analysis of the impact of signer overlap between the training and test sets on two sign detection benchmark datasets (Signing in the Wild (Borg and Camilleri, 2019) and the DGS Corpus (Hanke et al., 2020)) used by Borg and Camilleri (2019) and Moryossef et al. (2020). By comparing the accuracy with and without overlap, they noticed a relative decrease in performance for signers not present during training. As a result, they suggested new dataset partitions that eliminate overlap between train and test sets and facilitate a more accurate evaluation of performance.
2.2 Sign Language Segmentation
Segmentation consists of detecting the frame boundaries for signs or phrases in videos to divide them into meaningful units. While the most canonical way of dividing a spoken language text is into a linear sequence of words, due to the simultaneity of sign language, the notion of a sign language “word” is ill-defined, and sign language cannot be fully linearly modeled.
Current methods resort to segmenting units loosely mapped to signed language units (Santemiz et al., 2009; Farag and Brock, 2019; Bull et al., 2020b; Renz et al., 2021a, b; Bull et al., 2021) and do not explicitly leverage reliable linguistic predictors of sentence boundaries such as prosody in signed languages (i.e., pauses, extended sign duration, facial expressions) (Sandler, 2010; Ormel and Crasborn, 2012). De Sisto et al. (2021) call for a better understanding of sign language structure, which they believe is the necessary ground for the design and development of sign language recognition and segmentation methodologies.
Santemiz et al. (2009) automatically extracted isolated signs from continuous signing by aligning the sequences obtained via speech recognition, modeled by Dynamic Time Warping (DTW) and Hidden Markov Models (HMMs) approaches.
Farag and Brock (2019) used a random forest classifier to distinguish frames containing signs in Japanese Sign Language based on the composition of spatio-temporal angular and distance features between domain-specific pairs of joint segments.
Bull et al. (2020b) segmented French Sign Language into segments corresponding to subtitle units by relying on the alignment between subtitles and sign language videos, leveraging a spatio-temporal graph convolutional network (STGCN; Yu et al. (2017)) with a BiLSTM on 2D skeleton data.
Renz et al. (2021a) located temporal boundaries between signs in continuous sign language videos by employing 3D convolutional neural network representations with iterative temporal segment refinement to resolve ambiguities between sign boundary cues. Renz et al. (2021b) further proposed the Changepoint-Modulated Pseudo-Labelling (CMPL) algorithm to solve the problem of source-free domain adaptation.
3 Motivating Observations
To motivate our proposed approach, we make a series of observations regarding the intrinsic nature of sign language expressions. Specifically, we highlight the unique challenges posed by the continuous flow of sign language expressions (§3.1), the role of prosody in determining phrase boundaries (§3.2), and the influence of hand shape changes in indicating sign boundaries (§3.3).
3.1 Boundary Modeling
When examining the nature of sign language expressions, we note that the utterances are typically signed in a continuous flow, with minimal to no pauses between individual signs. This continuity is particularly evident when dealing with lower frame rates. This continuous nature presents a significant difference from text where specific punctuation marks serve as indicators of phrase boundaries, and a semi-closed set of tokens represent the words.
Given these characteristics, directly applying conventional segmentation or sign language detection models—that is, utilizing IO tagging in a manner similar to image or audio segmentation models—may not yield the optimal solution, particularly at the sign level. Such models often fail to precisely identify the boundaries between signs.
A promising alternative is the Beginning-Inside-Outside (BIO) tagging (Ramshaw and Marcus, 1995). BIO tagging was originally used for named entity recognition, but its application extends to any sequence chunking task beyond the text modality. In the context of sign language, BIO tagging provides a more refined model for discerning boundaries between signs and phrases, thus significantly improving segmentation performance (Figure 1).
To test the viability of the BIO tagging approach in comparison with the IO tagging, we conducted an experiment on the Public DGS Corpus. The entire corpus was transformed to various frame rates and the sign segments were converted to frames using either BIO or IO tagging, then decoded back into sign segments. Figure 4 illustrates the results of this comparison. Note that the IO tagging was unable to reproduce the same number of segments as the BIO tagging on the gold data. This underscores the importance of BIO tagging in identifying sign and phrase boundaries.

3.2 Phrase Boundaries
Linguistic research has shown that prosody is a reliable predictor of phrase boundaries in signed languages (Sandler, 2010; Ormel and Crasborn, 2012). We observe that this is also the case in the Public DGS Corpus dataset used in our experiments. To illustrate this, we model pauses and movement using optical flow directly on the poses as proposed by Moryossef et al. (2020). Figure 3 demonstrates that a change in motion signifies the presence of a pause, which, in turn, indicates a phrase boundary.
3.3 Sign Boundaries
We observe that signs generally utilize a limited number of hand shapes, with the majority of signs utilizing a maximum of two hand shapes. For example, linguistically annotated datasets, such as ASL-LEX Sehyr et al. (2021) and ASLLVD Neidle et al. (2012), only record one initial hand shape per hand and one final hand shape. Mandel (1981, p. 87) argued that there can only be one set of selected fingers per sign, constraining the number of handshapes in signs. This limitation is referred to as the Selected Fingers Constraint. And indeed, Sandler et al. (2008) find that the optimal form of a sign is monosyllabic, and that handshape change is organized by the syllable unit.
To illustrate this constraint empirically, we show a histogram of hand shapes per sign in SignBank222https://signbank.org/signpuddle2.0/ for signs in Figure 5.

Additionally, we found that a change in the dominant hand shape often signals the presence of a sign boundary. Specifically, out of sentences, comprising pairs of consecutive signs, only of consecutive signs share the same base hand shape333It is important to note that this percentage is inflated, as it may encompass overlaps across the dominant and non-dominant hands, which were not separated for this analysis.. Based on these observations, we propose using 3D hand normalization as an indicative cue for hand shapes to assist in detecting sign boundaries. We hypothesize that performing 3D hand normalization makes it easier for the model to extract the hand shape. We expand on this process and show examples in Appendix C.
4 Experimental Setup
In this section, we describe the experimental setup used to evaluate our linguistically motivated approach for sign language segmentation. This includes a description of the Public DGS Corpus dataset used in the study, the methodology employed to perform sign and phrase segmentation, and the evaluation metrics used to measure the performance of the proposed approach.
4.1 Dataset
The Public DGS Corpus Prillwitz et al. (2008); Hanke et al. (2020) is a distinctive sign language dataset that includes both accurate sign-level annotation from continuous signing, and well-aligned phrase-level translation in spoken language.
The corpus comprises 404 documents / 714 videos444The number of videos is nearly double the number of documents because each document typically includes two signers, each of whom produces one video for segmentation. with an average duration of 7.55 minutes, featuring either one signer or two signers, at 50 fps. Most of these videos feature gloss transcriptions and spoken language translations (German and English), except for the ones in the “Joke” category, which are not annotated and thus excluded from our model555We also exclude documents with missing annotations. . The translations are comprised of full spoken language paragraphs, sentences, or phrases (i.e., independent/main clauses).
Each gloss span is considered a gold sign segment, following a tight annotation scheme (Hanke et al., 2012). Phrase segments are identified by examining every translation, with the segment assumed to span from the start of its first sign to the end of its last sign, correcting imprecise annotation.
This corpus is enriched with full-body pose estimations from OpenPose Cao et al. (2019); Schulder and Hanke (2019) and Mediapipe Holistic Grishchenko and Bazarevsky (2020). We use the 3.0.0-uzh-document split from Zhang et al. (2023). After filtering the unannotated data, we are left with 296 documents / 583 videos for training, 6 / 12 for validation, and 9 / 17 for testing. The mean number of signs and phrases in a video from the training set is 613 and 111, respectively.
4.2 Methodology
Our proposed approach for sign language segmentation is based on the following steps:
-
1.
Pose Estimation Given a video, we first adjust it to 25 fps and estimate body poses using the MediaPipe Holistic pose estimation system. We do not use OpenPose because it lacks a -axis, which prevents 3D rotation used for hand normalization. The shape of a pose is represented as .
-
2.
Pose Normalization To generalize over video resolution and distance from the camera, we normalize each of these poses such that the mean distance between the shoulders of each person equals , and the mid-point is at (Celebi et al., 2013). We also remove the legs since they are less relevant to signing.
-
3.
Optical Flow We follow the equation in Moryossef et al. (2020, Equation 1).
-
4.
3D Hand Normalization We rotate and scale each hand to ensure that the same hand shape is represented in a consistent manner across different frames. We rotate the 21 keypoints of the hand so that the back of the hand lies on the plane, we then rotate the hand so that the metacarpal bone of the middle finger lies on the -axis, and finally, we scale the hand such that the bone is of constant length. Visualizations are presented in Appendix C.
-
5.
Sequence Encoder For every frame, the pose is first flattened and projected into a standard dimension (), then fed through an LSTM encoder (Hochreiter and Schmidhuber, 1997).
-
6.
BIO Tagging On top of the encoder, we place two BIO classification heads for sign and phrase independently. B denotes the beginning of a sign or phrase, I denotes the middle of a sign or phrase, and O denotes being outside a sign or phrase. Our cross-entropy loss is proportionally weighted in favor of B as it is a rare label666B:I:O is about 1:5:18 for signs and 1:58:77 for phrases. compared to I and O.
-
7.
Greedy Segment Decoding To decode the frame-level BIO predictions into sign/phrase segments, we define a segment to start with the first frame possessing a B probability greater than a predetermined threshold (defaulted at ). The segment concludes with the first frame among the subsequent frames, having either a B or O probability exceeding the threshold. We provide the pseudocode of the decoding algorithm in Appendix B.
4.3 Experiments
Our approach is evaluated through a series of six sets of experiments. Each set is repeated three times with varying random seeds. Preliminary experiments were conducted to inform the selection of hyperparameters and features, the details of which can be found in Table 3 in Appendix A. Model selection is based on validation metrics.
Sign | Phrase | Efficiency | |||||||
Experiment | F1 | IoU | % | F1 | IoU | % | #Params | Time | |
E0 | Moryossef et al. (2020) | — | — | 1.00 | 102K | 0:50:17 | |||
E1 | Baseline | 454K | 1:01:50 | ||||||
E2 | E1 + Face | 552K | 1:50:31 | ||||||
E3 | E1 + Optical Flow | 473K | 1:20:17 | ||||||
E4 | E3 + Hand Norm | 516K | 1:30:59 | ||||||
E1s | E1 + Depth=4 | 0.63 | 0.69 | 0.65 | 1.6M | 4:08:48 | |||
E2s | E2 + Depth=4 | 0.69 | 1.7M | 3:14:03 | |||||
E3s | E3 + Depth=4 | 1.7M | 4:08:30 | ||||||
E4s | E4 + Depth=4 | 1.7M | 4:35:29 | ||||||
E1s* | E1s + Tuned Decoding | — | 0.69 | 1.03 | — | 0.85 | 1.02 | — | — |
E4s* | E4s + Tuned Decoding | — | 0.63 | 1.06 | — | 0.79 | 1.12 | — | — |
E5 | E4s + Autoregressive | 1.3M | ~3 days |
-
1.
E0: IO Tagger We re-implemented and reproduced777The initial implementation uses OpenPose (Cao et al., 2019), at 50 fps. Preliminary experiments reveal that these differences do not significantly influence the results. the sign language detection model proposed by Moryossef et al. (2020), in PyTorch (Paszke et al., 2019) as a naive baseline. This model processes optical flow as input and outputs I (is signing) and O (not signing) tags.
-
2.
E1: Bidirectional BIO Tagger We replace the IO tagging heads in E0 with BIO heads to form our baseline. Our preliminary experiments indicate that inputting only the 75 hand and body keypoints and making the LSTM layer bidirectional yields optimal results.
-
3.
E2: Adding Reduced Face Keypoints Although the 75 hand and body keypoints serve as an efficient minimal set for sign language detection/segmentation models, we investigate the impact of other nonmanual sign language articulators, namely, the face. We introduce a reduced set of 128 face keypoints that signify the signer’s face contour888We reduce the dense FACE_LANDMARKS in Mediapipe Holistic to the contour keypoints according to the variable mediapipe.solutions.holistic.FACEMESH_CONTOURS..
-
4.
E3: Adding Optical Flow At every time step we append the optical flow between and to the current pose frame as an additional dimension after the axes.
-
5.
E4: Adding 3D Hand Normalization At every time step, we normalize the hand poses and concatenate them to the current pose.
-
6.
E5: Autoregressive Encoder We replace the encoder with the one proposed by Jiang et al. (2023) for the detection and classification of great ape calls from raw audio signals. Specifically, we add autoregressive connections between time steps to encourage consistent output labels. The logits at time step are concatenated to the input of the next time step, . This modification is implemented bidirectionally by stacking two autoregressive encoders and adding their output up before the Softmax operation. However, this approach is inherently slow, as we have to fully wait for the previous time step predictions before we can feed them to the next time step.
4.4 Evaluation Metrics
We evaluate the performance of our proposed approach for sign and phrase segmentation using the following metrics:
-
•
Frame-level F1 Score For each frame, we apply the argmax operation to make a local prediction of the BIO class and calculate the macro-averaged per-class F1 score against the ground truth label. We use this frame-level metric during validation as the primary metric for model selection and early stopping, due to its independence from a potentially variable segment decoding algorithm (§5.2).
-
•
Intersection over Union (IoU) We compute the IoU between the ground truth segments and the predicted segments to measure the degree of overlap. Note that we do not perform a one-to-one mapping between the two using techniques like DTW. Instead, we calculate the total IoU based on all segments in a video.
-
•
Percentage of Segments (%) To complement IoU, we introduce the percentage of segments to compare the number of segments predicted by the model with the ground truth annotations. It is computed as follows: . The optimal value is 1.
-
•
Efficiency We measure the efficiency of each model by the number of parameters and the training time of the model on a Tesla V100-SXM2-32GB GPU for 100 epochs999Exceptionally the autoregressive models in E5 were trained on an NVIDIA A100-SXM4-80GB GPUA100 which doubles the training speed of V100, still the training is slow..
5 Results and Discussion
We report the mean test evaluation metrics for our experiments in Table 1. We do not report F1 Score for E0 since it has a different number of classes and is thus incomparable. Comparing E1 to E0, we note that the model’s bidirectionality, the use of poses, and BIO tagging indeed help outperform the model from previous work where only optical flow and IO tagging are used. While E1 predicts an excessive number of phrase segments, the IoUs for signs and phrases are both higher.
Adding face keypoints (E2) makes the model worse, while including optical flow (E3) improves the F1 scores. For phrase segmentation, including optical flow increases IoU, but over-segments phrases by more than 300%, which further exaggerates the issue in E1. Including hand normalization (E4) on top of E3 slightly deteriorates the quality of both sign and phrase segmentation.
From the non-exhaustive hyperparameter search in the preliminary experiments (Table 3), we examined different hidden state sizes (, , , , ) and a range of to LSTM layers, and conclude that a hidden size of and layers with learning rate are optimal for E1, which lead to E1s. We repeat the setup of E2, E3, and E4 with these refined hyper-parameters, and show that all of them outperform their counterparts, notably in that they ease the phrase over-segmentation problem.
In E2s, we reaffirmed that adding face keypoints does not yield beneficial results, so we exclude face in future experiments. Although the face is an essential component to understanding sign language expressions and does play some role in sign and phrase level segmentation, we believe that the 128 face contour points are too dense for the model to learn useful information compared to the 75 body points, and may instead confuse the model.
In addition, the benefits of explicitly including optical flow (E3s) fade away with the increased model depth and we speculate that now the model might be able to learn the optical flow features by itself. Surprisingly, while adding hand normalization (E4s) still slightly worsens the overall results, it has the best phrase percentage.
From E4s we proceeded with the training of E5, an autoregressive model. Unexpectedly, counter to our intuition and previous work, E5 underachieves our baseline across all evaluation metrics101010E5 should have more parameters than E4s, but because of an implementation bug, each LSTM layer has half the parameters. Based on the current results, we assume that autoregressive connections (even with more parameters) will not improve our models..
5.1 Challenges with 3D Hand Normalization
While the use of 3D hand normalization is well-justified in §3.3, we believe it does not help the model due to poor depth estimation quality, as further corroborated by recent research from De Coster et al. (2023). Therefore, we consider it a negative result, showing the deficiencies in the 3D pose estimation system. The evaluation metrics we propose in Appendix C could help identify better pose estimation models for this use case.
5.2 Tuning the Segment Decoding Algorithm
We selected E1s and E4s to further explore the segment decoding algorithm. As detailed in §4.2 and Appendix B, the decoding algorithm has two tunable parameters, and . We conducted a grid search with these parameters, using values from 10 to 90 in increments of 10. We additionally experimented with a variation of the algorithm that conditions on the most likely class by argmax instead of fixed threshold values, which turned out similar to the default version.
We only measured the results using IoU and the percentage of segments at validation time since the F1 scores remain consistent in this case. For sign segmentation, we found using and yields slightly better results than the default setting ( for both). For phrase segmentation, we identified that higher threshold values ( for E1s and for E4s) improve on the default significantly, especially on the percentage metric. We report the test results under E1s* and E4s*, respectively.
Despite formulating a single model, we underline a separate sign/phrase model selection process to archive the best segmentation results. Figure 6 illustrates how higher threshold values reduce the number of predicted segments and skew the distribution of predicted phrase segments towards longer ones in E1s/E1s*. As Bull et al. (2020b) suggest, advanced priors could also be integrated into the decoding algorithm.

5.3 Comparison to Previous Work
We re-implemented and re-purposed the sign language detection model introduced in Moryossef et al. (2020) for our segmentation task as a baseline since their work is the state-of-the-art and the only comparable model designed for the Public DGS Corpus dataset. As a result, we show the necessity of replacing IO tagging with BIO tagging to tackle the subtle differences between the two tasks.
For phrase segmentation, we compare to Bull et al. (2020b). We note that our definition of sign language phrases (spanning from the start of its first sign to the end of its last sign) is tighter than the subtitle units used in their paper and that we use different training datasets of different languages and domains. Nevertheless, we implemented some of their frame-level metrics and show the results in Table 2 on both the Public DGS Corpus and the MEDIAPI-SKEL dataset (Bull et al., 2020a) in French Sign Language (LSF). We report both zero-shot out-of-domain results111111The zero-shot results are not directly comparable to theirs due to different datasets and labeling approaches. and the results of our models trained specifically on their dataset without the spatio-temporal graph convolutional network (ST-GCN) (Yan et al., 2018) used in their work for pose encoding.
Data | Model | ROC-AUC | F1-M |
---|---|---|---|
LSF | full (theirs) | 0.87 | — |
body (theirs) | 0.87 | — | |
E1s (ours, zero-shot) | 0.71 | 0.41 | |
E4s (ours, zero-shot) | 0.76 | 0.44 | |
E1s (ours, trained) | 0.87 | 0.49 | |
E4s (ours, trained) | 0.87 | 0.51 | |
DGS | E1s (ours) | 0.91 | 0.65 |
E4s (ours) | 0.90 | 0.62 |
6 Conclusions
This work focuses on the automatic segmentation of signed languages. We are the first to formulate the segmentation of individual signs and larger sign phrases as a joint problem.
We propose a series of improvements over previous work, linguistically motivated by careful analyses of sign language corpora. Recognizing that sign language utterances are typically continuous with minimal pauses, we opted for a BIO tagging scheme over IO tagging. Furthermore, leveraging the fact that phrase boundaries are marked by prosodic cues, we introduce optical flow features as a proxy for prosodic processes. Finally, since signs typically employ a limited number of hand shapes, to make it easier for the model to understand handshapes, we attempt 3D hand normalization.
Our experiments conducted on the Public DGS Corpus confirmed the efficacy of these modifications for segmentation quality. By comparing to previous work in a zero-shot setting, we demonstrate that our models generalize across signed languages and domains and that including linguistically motivated cues leads to a more robust model in this context.
Finally, we envision that the proposed model has applications in real-world data collection for signed languages. Furthermore, a similar segmentation approach could be leveraged in various other fields such as co-speech gesture recognition (Moryossef, 2023) and action segmentation (Tang et al., 2019).
Limitations
Pose Estimation
In this work, we employ the MediaPipe Holistic pose estimation system (Grishchenko and Bazarevsky, 2020). There is a possibility that this system exhibits bias towards certain protected classes (such as gender or race), underperforming in instances with specific skin tones or lower video quality. Thus, we cannot attest to how our system would perform under real-world conditions, given that the videos utilized in our research are generated in a controlled studio environment, primarily featuring white participants.
Encoding of Long Sequences
In this study, we encode sequences of frames that are significantly longer than the typical 512 frames often seen in models employing Transformers (Vaswani et al., 2017). Numerous techniques, ranging from basic temporal pooling/downsampling to more advanced methods such as a video/pose encoder that aggregates local frames into higher-level ‘tokens’ (Renz et al., 2021a), graph convolutional networks (Bull et al., 2020b), and self-supervised representations (Baevski et al., 2020), can alleviate length constraints, facilitate the use of Transformers, and potentially improve the outcomes. Moreover, a hierarchical method like the Swin Transformer (Liu et al., 2021) could be applicable.
Limitations of Autoregressive LSTMs
In this paper, we replicated the autoregressive LSTM implementation originally proposed by Jiang et al. (2023). Our experiments revealed that this implementation exhibits significant slowness, which prevented us from performing further experimentation. In contrast, other LSTM implementations employed in this project have undergone extensive optimization (Appleyard, 2016), including techniques like combining general matrix multiplication operations (GEMMs), parallelizing independent operations, fusing kernels, rearranging matrices, and implementing various optimizations for models with multiple layers (which are not necessarily applicable here).
A comparison of CPU-based performance demonstrates that our implementation is x6.4 times slower. Theoretically, the number of operations performed by the autoregressive LSTM is equivalent to that of a regular LSTM. However, while the normal LSTM benefits from concurrency based on the number of layers, we do not have that luxury. The optimization of recurrent neural networks (RNNs) (Que et al., 2020, 2021, 2022) remains an ongoing area of research. If proven effective in other domains, we strongly advocate for efforts to optimize the performance of this type of network.
Interference Between Sign and Phrase Models
In our model, we share the encoder for both the sign and phrase segmentation models, with a shallow linear layer for the BIO tag prediction associated with each task. It remains uncertain whether these two tasks interfere with or enhance each other. An ablation study (not presented in this work) involving separate modeling is necessary to obtain greater insight into this matter.
Noisy Training Objective
Although the annotations utilized in this study are of expert level, the determination of precise sign (Hanke et al., 2012) and phrase boundaries remains a challenging task, even for experts. Training the model on these annotated boundaries might introduce excessive noise. A similar issue was observed in classification-based pose estimation (Cao et al., 2019). The task of annotating the exact anatomical centers of joints proves to be nearly impossible, leading to a high degree of noise when predicting joint position as a 1-hot classification task. The solution proposed in this previous work was to distribute a Gaussian around the annotated location of each joint. This approach allows the joint’s center to overlap with some probability mass, thereby reducing the noise for the model. A similar solution could be applied in our context. Instead of predicting a strict 0 or 1 class probability, we could distribute a Gaussian around the boundary.
Naive Segment Decoding
We recognize that the frame-level greedy decoding strategy implemented in our study may not be optimal. Previous research in audio segmentation (Venkatesh et al., 2022) employed a You Only Look Once (YOLO; Redmon et al. (2015)) decoding scheme to predict segment boundaries and classes. We propose using a similar prediction atop a given representation, such as the LSTM output or classification logits of an already trained network. Differing from traditional object detection tasks, this process is simplified due to the absence of a axis and non-overlapping segments. In this scenario, the network predicts the segment boundaries using regression, thereby avoiding the class imbalance issue of the BIO tagging. We anticipate this to yield more accurate sign language segmentation.
Lack of Transcription
Speech segmentation is a close task to our sign language segmentation task on videos. In addition to relying on prosodic cues from audio, the former could benefit from automatic speech transcription systems, either in terms of surrogating the task to text-level segmentation and punctuation (Cho et al., 2015), or gaining additional training data from automatic speech recognition / spoken language translation (Tsiamas et al., 2022).
However, for signed languages, there is neither a standardized and widely used written form nor a reliable transcription procedure into some potential writing systems like SignWriting (Sutton, 1990), HamNoSys (Prillwitz and Zienert, 1990), and glosses (Johnston, 2008). Transcription/recognition and segmentation tasks need to be solved simultaneously, so we envision that a multi-task setting helps. Sign spotting, the localization of a specific sign in continuous signing, is a simplification of the segmentation and recognition problem in a closed-vocabulary setting (Wong et al., 2022; Varol et al., 2022). It can be used to find candidate boundaries for some signs, but not all.
Acknowledgements
This work was funded by the EU Horizon 2020 project EASIER (grant agreement no. 101016982), the Swiss Innovation Agency (Innosuisse) flagship IICT (PFFS-21-47) and the EU Horizon 2020 project iEXTRACT (grant agreement no. 802774). We also thank Rico Sennrich and Chantal Amrhein for their suggestions.
References
- Appleyard (2016) Jeremy Appleyard. 2016. Optimizing recurrent neural networks in cuDNN 5. Accessed: 2023-06-09.
- Baevski et al. (2020) Alexei Baevski, Yuhao Zhou, Abdelrahman Mohamed, and Michael Auli. 2020. wav2vec 2.0: A framework for self-supervised learning of speech representations. Advances in neural information processing systems, 33:12449–12460.
- Borg and Camilleri (2019) Mark Borg and Kenneth P Camilleri. 2019. Sign language detection "in the wild" with recurrent neural networks. In ICASSP 2019-2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pages 1637–1641. IEEE.
- Bull et al. (2021) Hannah Bull, Triantafyllos Afouras, Gül Varol, Samuel Albanie, Liliane Momeni, and Andrew Zisserman. 2021. Aligning subtitles in sign language videos. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 11552–11561.
- Bull et al. (2020a) Hannah Bull, Annelies Braffort, and Michèle Gouiffès. 2020a. MEDIAPI-SKEL - a 2D-skeleton video database of French Sign Language with aligned French subtitles. In Proceedings of the Twelfth Language Resources and Evaluation Conference, pages 6063–6068, Marseille, France. European Language Resources Association.
- Bull et al. (2020b) Hannah Bull, Michèle Gouiffès, and Annelies Braffort. 2020b. Automatic segmentation of sign language into subtitle-units. In European Conference on Computer Vision, pages 186–198. Springer.
- Cao et al. (2019) Z. Cao, G. Hidalgo Martinez, T. Simon, S. Wei, and Y. A. Sheikh. 2019. OpenPose: Realtime multi-person 2D pose estimation using part affinity fields. IEEE Transactions on Pattern Analysis and Machine Intelligence.
- Celebi et al. (2013) Sait Celebi, Ali Selman Aydin, Talha Tarik Temiz, and Tarik Arici. 2013. Gesture recognition using skeleton data with weighted dynamic time warping. In International Conference on Computer Vision Theory and Applications.
- Cho et al. (2015) Eunah Cho, Jan Niehues, Kevin Kilgour, and Alex Waibel. 2015. Punctuation insertion for real-time spoken language translation. In Proceedings of the 12th International Workshop on Spoken Language Translation: Papers, pages 173–179.
- Cho et al. (2014) Kyunghyun Cho, Bart van Merriënboer, Caglar Gulcehre, Dzmitry Bahdanau, Fethi Bougares, Holger Schwenk, and Yoshua Bengio. 2014. Learning phrase representations using RNN encoder–decoder for statistical machine translation. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 1724–1734, Doha, Qatar. Association for Computational Linguistics.
- De Coster et al. (2023) Mathieu De Coster, Ellen Rushe, Ruth Holmes, Anthony Ventresque, and Joni Dambre. 2023. Towards the extraction of robust sign embeddings for low resource sign language recognition. arXiv preprint arXiv:2306.17558.
- De Sisto et al. (2021) Mirella De Sisto, Dimitar Shterionov, Irene Murtagh, Myriam Vermeerbergen, and Lorraine Leeson. 2021. Defining meaningful units. challenges in sign segmentation and segment-meaning mapping (short paper). In Proceedings of the 1st International Workshop on Automatic Translation for Signed and Spoken Languages (AT4SSL), pages 98–103, Virtual. Association for Machine Translation in the Americas.
- Devlin et al. (2019) Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2019. BERT: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), pages 4171–4186, Minneapolis, Minnesota. Association for Computational Linguistics.
- Farag and Brock (2019) Iva Farag and Heike Brock. 2019. Learning motion disfluencies for automatic sign language segmentation. In ICASSP 2019-2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pages 7360–7364. IEEE.
- Frost and Sutton (2022) Adam Frost and Valerie Sutton. 2022. SignWriting Hand Symbols. SignWriting Press.
- Grishchenko and Bazarevsky (2020) Ivan Grishchenko and Valentin Bazarevsky. 2020. Mediapipe holistic.
- Hanke et al. (2012) Thomas Hanke, Silke Matthes, Anja Regen, and Satu Worseck. 2012. Where does a sign start and end? segmentation of continuous signing. In Language Resources and Evaluation Conference.
- Hanke et al. (2020) Thomas Hanke, Marc Schulder, Reiner Konrad, and Elena Jahn. 2020. Extending the Public DGS Corpus in size and depth. In Proceedings of the LREC2020 9th Workshop on the Representation and Processing of Sign Languages: Sign Language Resources in the Service of the Language Community, Technological Challenges and Application Perspectives, pages 75–82, Marseille, France. European Language Resources Association (ELRA).
- Hochreiter and Schmidhuber (1997) Sepp Hochreiter and Jürgen Schmidhuber. 1997. Long short-term memory. Neural computation, 9(8):1735–1780.
- Jiang et al. (2023) Zifan Jiang, Adrian Soldati, Isaac Schamberg, Adriano R Lameira, and Steven Moran. 2023. Automatic sound event detection and classification of great ape calls using neural networks. arXiv preprint arXiv:2301.02214.
- Johnston (2008) Trevor Alexander Johnston. 2008. From archive to corpus: transcription and annotation in the creation of signed language corpora. In Pacific Asia Conference on Language, Information and Computation.
- Liu et al. (2021) Ze Liu, Yutong Lin, Yue Cao, Han Hu, Yixuan Wei, Zheng Zhang, Stephen Lin, and Baining Guo. 2021. Swin transformer: Hierarchical vision transformer using shifted windows. In Proceedings of the IEEE/CVF international conference on computer vision, pages 10012–10022.
- Mandel (1981) Mark Alan Mandel. 1981. Phonotactics and morphophonology in American Sign Language. University of California, Berkeley.
- Moryossef (2023) Amit Moryossef. 2023. Addressing the blind spots in spoken language processing. arXiv preprint arXiv:2309.06572.
- Moryossef et al. (2020) Amit Moryossef, Ioannis Tsochantaridis, Roee Aharoni, Sarah Ebling, and Srini Narayanan. 2020. Real-time sign-language detection using human pose estimation. In Computer Vision–ECCV 2020 Workshops: Glasgow, UK, August 23–28, 2020, Proceedings, Part II 16, SLRTP 2020: The Sign Language Recognition, Translation and Production Workshop, pages 237–248.
- Neidle et al. (2012) Carol Neidle, Ashwin Thangali, and Stan Sclaroff. 2012. Challenges in development of the American Sign Language lexicon video dataset (ASSLVD) corpus. In 5th workshop on the representation and processing of sign languages: interactions between corpus and Lexicon, LREC. Citeseer.
- Ormel and Crasborn (2012) Ellen Ormel and Onno Crasborn. 2012. Prosodic correlates of sentences in signed languages: A literature review and suggestions for new types of studies. Sign Language Studies, 12(2):279–315.
- Pal et al. (2023) Abhilash Pal, Stephan Huber, Cyrine Chaabani, Alessandro Manzotti, and Oscar Koller. 2023. On the importance of signer overlap for sign language detection. arXiv preprint arXiv:2303.10782.
- Paszke et al. (2019) Adam Paszke, Sam Gross, Francisco Massa, Adam Lerer, James Bradbury, Gregory Chanan, Trevor Killeen, Zeming Lin, Natalia Gimelshein, Luca Antiga, Alban Desmaison, Andreas Kopf, Edward Yang, Zachary DeVito, Martin Raison, Alykhan Tejani, Sasank Chilamkurthy, Benoit Steiner, Lu Fang, Junjie Bai, and Soumith Chintala. 2019. PyTorch: An imperative style, high-performance deep learning library. In Advances in Neural Information Processing Systems 32, pages 8024–8035. Curran Associates, Inc.
- Prillwitz et al. (2008) Siegmund Prillwitz, Thomas Hanke, Susanne König, Reiner Konrad, Gabriele Langer, and Arvid Schwarz. 2008. Dgs corpus project–development of a corpus based electronic dictionary german sign language/german. In sign-lang at LREC 2008, pages 159–164. European Language Resources Association (ELRA).
- Prillwitz and Zienert (1990) Siegmund Prillwitz and Heiko Zienert. 1990. Hamburg notation system for sign language: Development of a sign writing with computer application. In Current trends in European Sign Language Research. Proceedings of the 3rd European Congress on Sign Language Research, pages 355–379.
- Que et al. (2022) Zhiqiang Que, Hiroki Nakahara, Hongxiang Fan, He Li, Jiuxi Meng, Kuen Hung Tsoi, Xinyu Niu, Eriko Nurvitadhi, and Wayne W. C. Luk. 2022. Remarn: A reconfigurable multi-threaded multi-core accelerator for recurrent neural networks. ACM Transactions on Reconfigurable Technology and Systems, 16:1 – 26.
- Que et al. (2020) Zhiqiang Que, Hiroki Nakahara, Eriko Nurvitadhi, Hongxiang Fan, Chenglong Zeng, Jiuxi Meng, Xinyu Niu, and Wayne W. C. Luk. 2020. Optimizing reconfigurable recurrent neural networks. 2020 IEEE 28th Annual International Symposium on Field-Programmable Custom Computing Machines (FCCM), pages 10–18.
- Que et al. (2021) Zhiqiang Que, Erwei Wang, Umar Marikar, Eric Moreno, Jennifer Ngadiuba, Hamza Javed, Bartłomiej Borzyszkowski, Thea Klaeboe Aarrestad, Vladimir Loncar, Sioni Paris Summers, Maurizio Pierini, Peter Y. K. Cheung, and Wayne W. C. Luk. 2021. Accelerating recurrent neural networks for gravitational wave experiments. 2021 IEEE 32nd International Conference on Application-specific Systems, Architectures and Processors (ASAP), pages 117–124.
- Ramırez et al. (2004) Javier Ramırez, José C Segura, Carmen Benıtez, Angel De La Torre, and Antonio Rubio. 2004. Efficient voice activity detection algorithms using long-term speech information. Speech communication, 42(3-4):271–287.
- Ramshaw and Marcus (1995) Lance Ramshaw and Mitch Marcus. 1995. Text chunking using transformation-based learning. In Third Workshop on Very Large Corpora.
- Redmon et al. (2015) Joseph Redmon, Santosh Kumar Divvala, Ross B. Girshick, and Ali Farhadi. 2015. You only look once: Unified, real-time object detection. 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 779–788.
- Renz et al. (2021a) Katrin Renz, Nicolaj C Stache, Samuel Albanie, and Gül Varol. 2021a. Sign language segmentation with temporal convolutional networks. In ICASSP 2021-2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pages 2135–2139. IEEE.
- Renz et al. (2021b) Katrin Renz, Nicolaj C Stache, Neil Fox, Gul Varol, and Samuel Albanie. 2021b. Sign segmentation with changepoint-modulated pseudo-labelling. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 3403–3412.
- Sandler (2010) Wendy Sandler. 2010. Prosody and syntax in sign languages. Transactions of the philological society, 108(3):298–328.
- Sandler and Lillo-Martin (2006) Wendy Sandler and Diane Lillo-Martin. 2006. Sign language and linguistic universals. Cambridge University Press.
- Sandler et al. (2008) Wendy Sandler, Peter Macneilage, Barbara Davis, Kristine Zajdo, New York, and Taylor Francis. 2008. The syllable in sign language: Considering the other natural language modality.
- Santemiz et al. (2009) Pinar Santemiz, Oya Aran, Murat Saraclar, and Lale Akarun. 2009. Automatic sign segmentation from continuous signing via multiple sequence alignment. In 2009 IEEE 12th International Conference on Computer Vision Workshops, ICCV Workshops, pages 2001–2008. IEEE.
- Schulder and Hanke (2019) Marc Schulder and Thomas Hanke. 2019. OpenPose in the Public DGS Corpus. Project Note AP06-2019-01, DGS-Korpus project, IDGS, Hamburg University, Hamburg, Germany.
- Sehyr et al. (2021) Zed Sevcikova Sehyr, Naomi Caselli, Ariel M Cohen-Goldberg, and Karen Emmorey. 2021. The ASL-LEX 2.0 project: A database of lexical and phonological properties for 2,723 signs in american sign language. The Journal of Deaf Studies and Deaf Education, 26(2):263–277.
- Simonyan and Zisserman (2015) Karen Simonyan and Andrew Zisserman. 2015. Very deep convolutional networks for large-scale image recognition. CoRR.
- Sohn et al. (1999) Jongseo Sohn, Nam Soo Kim, and Wonyong Sung. 1999. A statistical model-based voice activity detection. IEEE signal processing letters, 6(1):1–3.
- Stokoe Jr (1960) William C Stokoe Jr. 1960. Sign Language Structure: An Outline of the Visual Communication Systems of the American Deaf. The Journal of Deaf Studies and Deaf Education, 10(1):3–37.
- Sutton (1990) Valerie Sutton. 1990. Lessons in sign writing. SignWriting.
- Tang et al. (2019) Yansong Tang, Dajun Ding, Yongming Rao, Yu Zheng, Danyang Zhang, Lili Zhao, Jiwen Lu, and Jie Zhou. 2019. COIN: A large-scale dataset for comprehensive instructional video analysis. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
- Tsiamas et al. (2022) Ioannis Tsiamas, Gerard I. Gállego, José A. R. Fonollosa, and Marta R. Costa-jussà. 2022. SHAS: Approaching optimal segmentation for end-to-end speech translation. In Proc. Interspeech 2022, pages 106–110.
- Varol et al. (2022) Gül Varol, Liliane Momeni, Samuel Albanie, Triantafyllos Afouras, and Andrew Zisserman. 2022. Scaling up sign spotting through sign language dictionaries. International Journal of Computer Vision, 130:1416 – 1439.
- Vaswani et al. (2017) Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, Łukasz Kaiser, and Illia Polosukhin. 2017. Attention is all you need. Advances in neural information processing systems, 30.
- Venkatesh et al. (2022) Satvik Venkatesh, David Moffat, and Eduardo Reck Miranda. 2022. You only hear once: a YOLO-like algorithm for audio segmentation and sound event detection. Applied Sciences, 12(7):3293.
- Wong et al. (2022) Ryan Wong, Necati Cihan Camgoz, and R. Bowden. 2022. Hierarchical I3D for sign spotting. In ECCV Workshops.
- Yan et al. (2018) Sijie Yan, Yuanjun Xiong, and Dahua Lin. 2018. Spatial temporal graph convolutional networks for skeleton-based action recognition. In Proceedings of the AAAI conference on artificial intelligence, volume 32.
- Yu et al. (2017) Ting Yu, Haoteng Yin, and Zhanxing Zhu. 2017. Spatio-temporal graph convolutional networks: A deep learning framework for traffic forecasting. In International Joint Conference on Artificial Intelligence.
- Zhang et al. (2023) Biao Zhang, Mathias Müller, and Rico Sennrich. 2023. SLTUNET: A simple unified model for sign language translation. In The Eleventh International Conference on Learning Representations, Kigali, Rwanda.
Appendix A Extended Experimental Results
We conducted some preliminary experiments (starting with P0) on training a sign language segmentation model to gain insights into hyperparameters and feature choices. The results are shown in Table 3121212Note that due to an implementation issue on edge cases (which we fixed later), the IoU and % values in Table 3 are lower than the ones in Table 1 and Table 4 thus not comparable across tables. The comparison inside of Table 3 between different experiments remains meaningful. In addition, the results in Table 3 are based on only one run instead of three random runs.. We found in P1.3.2 the optimal hyperparameters and repeated them with different feature choices.
Sign | Phrase | |||||||
Experiment | F1 | IoU | % | F1 | IoU | % | ||
P0 | Moryossef et al. (2020) | test | — | — | ||||
dev | — | — | ||||||
P0.1 | P0 + Holistic 25fps | test | — | — | ||||
dev | — | — | ||||||
P1 | P1 baseline | test | ||||||
dev | ||||||||
P1.1 | P1 - encoder_bidirectional | test | ||||||
dev | ||||||||
P1.2.1 | P1 + hidden_dim=512 | test | ||||||
dev | ||||||||
P1.2.2 | P1 + hidden_dim=1024 | test | ||||||
dev | ||||||||
P1.3.1 | P1 + encoder_depth=2 | test | ||||||
dev | ||||||||
P1.3.2 | P1 + encoder_depth=4 | test | ||||||
dev | ||||||||
P1.4.1 | P1 + hidden_dim=128 + encoder_depth=2 | test | ||||||
dev | ||||||||
P1.4.2 | P1 + hidden_dim=128 + encoder_depth=4 | test | ||||||
dev | ||||||||
P1.4.3 | P1 + hidden_dim=128 + encoder_depth=8 | test | ||||||
dev | ||||||||
P1.5.1 | P1 + hidden_dim=64 + encoder_depth=4 | test | ||||||
dev | ||||||||
P1.5.2 | P1 + hidden_dim=64 + encoder_depth=8 | test | ||||||
dev | ||||||||
P2 | P1 + optical_flow | test | ||||||
dev | ||||||||
P2.1 | P1.3.2 + optical_flow | test | ||||||
dev | ||||||||
P3 | P1 + hand_normalization | test | ||||||
dev | ||||||||
P3.1 | P1.3.2 + hand_normalization | test | ||||||
dev | ||||||||
P4 | P2.1 + P3.1 | test | ||||||
dev | ||||||||
P4.1 | P4 + encoder_depth=8 | test | ||||||
dev | ||||||||
P5 | P1.3.2 + reduced_face | test | ||||||
dev | ||||||||
P5.1 | P1.3.2 + full_face | test | ||||||
dev |
We selected some promising models from our preliminary experiments and reran them three times using different random seeds to make the final conclusion reliable and robust. Table 4 includes the standard deviation and the validation results (where we performed the model selection) for readers to scrutinize.
Sign | Phrase | Efficiency | ||||||||
Experiment | F1 | IoU | % | F1 | IoU | % | #Params | Time | ||
E0 | Moryossef et al. (2020) | test | — | — | 102K | 0:50:17 | ||||
dev | — | — | 102K | 0:50:17 | ||||||
E1 | Baseline | test | 454K | 1:01:50 | ||||||
dev | 454K | 1:01:50 | ||||||||
E2 | E1 + Face | test | 552K | 1:50:31 | ||||||
dev | 552K | 1:50:31 | ||||||||
E3 | E1 + Optical Flow | test | 473K | 1:20:17 | ||||||
dev | 473K | 1:20:17 | ||||||||
E4 | E3 + Hand Norm | test | 516K | 1:30:59 | ||||||
dev | 516K | 1:30:59 | ||||||||
E1s | E1 + Depth=4 | test | 1.6M | 4:08:48 | ||||||
dev | 1.6M | 4:08:48 | ||||||||
E2s | E2 + Depth=4 | test | 1.7M | 3:14:03 | ||||||
dev | 1.7M | 3:14:03 | ||||||||
E3s | E3 + Depth=4 | test | 1.7M | 4:08:30 | ||||||
dev | 1.7M | 4:08:30 | ||||||||
E4s | E4 + Depth=4 | test | 1.7M | 4:35:29 | ||||||
dev | 1.7M | 4:35:29 | ||||||||
E4ba | E4s + Autoregressive | test | 1.3M | 2 days, 21:28:42 | ||||||
dev | 1.3M | 2 days, 21:28:42 |
Appendix B Greedy Decoding Algorithm
We provide our exact decoding algorithm in Algorithm 1. We opt to employ adjustable thresholds rather than argmax prediction, as our empirical findings demonstrate superior performance with this approach (§5.2).
Appendix C Pose Based Hand Shape Analysis
C.1 Introduction to Hand Shapes in Sign Language
The most prominent feature of signed languages is their use of the hands. In fact, the hands play an integral role in the phonetics of signs, and a slight variation in hand shape can convey differences in meaning Stokoe Jr (1960). In sign languages such as American Sign Language (ASL) and British Sign Language (BSL), different hand shapes contribute to the vocabulary of the language, similar to how different sounds contribute to the vocabulary of spoken languages. ASL is estimated to use between 30 to 80 hand shapes131313https://aslfont.github.io/Symbol-Font-For-ASL/asl/handshapes.html, while BSL is limited to approximately 40 hand shapes141414https://bsl.surrey.ac.uk/principles/i-hand-shapes. SignWriting Sutton (1990), a system of notation used for sign languages, specifies a superset of 261 distinct hand shapes Frost and Sutton (2022). Each sign language uses a subset of these hand shapes.
Despite the fundamental role of hand shapes in sign languages, accurately recognizing and classifying them is a challenging task. In this section, we explore rule-based hand shape analysis in sign languages using 3D hand normalization. By performing 3D hand normalization, we can transform any given hand shape to a fixed orientation, making it easier for a model to extract the hand shape, and hence improving the recognition and classification of hand shapes in sign languages.
C.2 Characteristics of the Human Hand
The human hand consists of 27 bones and can be divided into three main sections: the wrist (carpals), the palm (metacarpals), and the fingers (phalanges). Each finger consists of three bones, except for the thumb, which has two. The bones are connected by joints, which allow for the complex movements and shapes that the hand can form.

Understanding the different characteristics of hands and their implications in signed languages is crucial for the extraction and classification of hand shapes. These characteristics are based on the SignWriting definitions of the five major axes of hand variation: handedness, plane, rotation, view, and shape.
Handedness
is the distinction between the right and left hands. Signed languages make a distinction between the dominant hand and the non-dominant hand. For right-handed individuals, the right hand is considered dominant, and vice-versa. The dominant hand is used for fingerspelling and all one-handed signs, while the non-dominant hand is used for support and two-handed signs. Using 3D pose estimation, the handedness analysis is trivial, as the pose estimation platform predicts which hand is which.
Plane
refers to whether the hand is parallel to the wall or the floor. The variation in the plane can, but does not have to, create a distinction between two signs. For example, in ASL the signs for “date” and “dessert” exhibit the same hand shape, view, rotation, and movement, but differ by plane. The plane of a hand can be estimated by comparing the positions of the wrist and middle finger metacarpal bone ().
Rotation
refers to the angle of the hand in relation to the body. SignWriting groups the hand rotation into eight equal categories, each spanning 45 degrees. The rotation of a hand can be calculated by finding the angle of the line created by the wrist and the middle finger metacarpal bone.
View
refers to the side of the hand as observed by the signer, and is grouped into four categories: front, back, sideways, and other-sideways. The view of a hand can be estimated by analyzing the normal of the plane created by the palm of the hand (between the wrist, index finger metacarpal bone, and pinky metacarpal bone).
Shape
refers to the configuration of the fingers and thumb. This characteristic of the hand is the most complex to analyze due to the vast array of possible shapes the human hand can form. The shape of a hand is determined by the state of each finger and thumb, specifically whether they are straight, curved, or bent, and their position relative to each other. Shape analysis can be accomplished by examining the bend and rotation of each finger joint. More advanced models may also take into consideration the spread between the fingers and other nuanced characteristics. 3D pose estimation can be used to extract these features for a machine learning model, which can then classify the hand shape.
C.3 3D Hand Normalization
3D hand normalization is an attempt at standardizing the orientation and position of the hand, thereby enabling models to effectively classify various hand shapes. The normalization process involves several key steps, as illustrated below:
-
1.
Pose Estimation Initially, the 3D pose of the hand is estimated from the hand image crop (Figure 8).
Figure 8: Pictures of six hands all performing the same hand shape (v-shape) taken from six different orientations. Mediapipe fails at estimating the pose of the bottom-middle image. -
2.
3D Rotation The pose is then rotated in 3D space such that the normal of the back of the hand aligns with the -axis. As a result, the palm plane now resides within the plane (Figure 9).
Figure 9: Hand poses after 3D rotation. The scale difference between the hands demonstrates a limitation of the 3D pose estimation system used. -
3.
2D Orientation Subsequently, the pose is rotated in 2D such that the metacarpal bone of the middle finger aligns with the -axis (Figure 10).
Figure 10: Hand poses after being rotated. -
4.
Scale The hand is scaled such that the metacarpal bone of the middle finger attains a constant length (which we typically set to , Figure 11).
Figure 11: Hand poses after being scaled. -
5.
Translation Lastly, the wrist joint is translated to the origin of the coordinate system . Figure 12 demonstrates how when overlayed, we can see all hands producing the same shape, except for one outlier.
Figure 12: Normalized hand poses overlayed after being translated to the same position. The positions of the wrist and the metacarpal bone of the middle finger are fixed.
By conducting these normalization steps, a hand model can be standardized, reducing the complexity of subsequent steps such as feature extraction and hand shape classification. This standardization simplifies the recognition process and can contribute to improving the overall accuracy of the system.
C.4 3D Hand Pose Evaluation
In order to assess the performance of our 3D hand pose estimation and normalization, we introduce two metrics that gauge the consistency of the pose estimation across orientations and crops.
Our dataset is extracted from the SignWriting Hand Symbols Manual Frost and Sutton (2022), and includes images of 261 different hand shapes, from 6 different angles. All images are of the same hand, of an adult white man.
Multi Angle Consistency Error (MACE)
evaluates the consistency of the pose estimation system across the different orientations. We perform 3D hand normalization, and overlay the hands. The MACE score is the average standard deviation of all pose landmarks, between all views. A high MACE score indicates a problem in the pose estimation system’s ability to maintain consistency across different orientations. This could adversely affect the model’s performance when analyzing hand shapes in sign languages, as signs can significantly vary with hand rotation.










Figure 13 shows that our 3D normalization does work to some extent using Mediapipe. We can identify differences across hand shapes, but still note high variance within each hand shape.
Crop Consistency Error (CCE)
gauges the pose estimation system’s consistency across different crop sizes. We do not perform 3D normalization, but still overlay all the estimated hands, shifting the wrist point of each estimated hand to the origin . The CCE score is the calculated average standard deviation of all pose landmarks across crops. A high CCE score indicates that the pose estimation system is sensitive to the size of the input crop, which is a significant drawback as the system should be invariant to the size of the input image.










Figure 14 shows that for some poses, Mediapipe is very resilient to crop size differences (e.g. the first and last hand shapes). However, it is concerning that for some hand shapes, it exhibits very high variance, and possibly even wrong predictions.
C.5 Conclusion
Our normalization process appears to work reasonably well when applied to different views within the same crop size. It succeeds in simplifying the hand shape, which in turn, can aid in improving the accuracy of hand shape classification systems.
However, it is crucial to note that while this method may seem to perform well on a static image, its consistency and reliability in a dynamic context, such as a video, may be quite different. In a video, the crop size can change between frames, introducing additional complexity and variance. This dynamic nature coupled with the inherently noisy nature of the estimation process can pose challenges for a model that aims to consistently estimate hand shapes.
In light of these findings, it is clear that there is a need for the developers of 3D pose estimation systems to consider these evaluation methods and strive to make their systems more robust to changes in hand crops. The Multi Angle Consistency Error (MACE) and the Crop Consistency Error (CCE) can be valuable tools in this regard.
MACE could potentially be incorporated as a loss function for 3D pose estimation, thereby driving the model to maintain consistency across different orientations. Alternatively, MACE could be used as an indicator to identify hand shapes that require more training data. It is apparent from our study that the performance varies greatly across hand shapes and orientations, and this approach could help in prioritizing the allocation of training resources.
Ultimately, the goal of improving 3D hand pose estimation is to enhance the ability to encode signed languages accurately. The insights gathered from this study can guide future research and development efforts in this direction, paving the way for more robust and reliable sign language technology.
The benchmark, metrics, and visualizations are available at https://github.com/sign-language-processing/3d-hands-benchmark/.