This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Including Facial Expressions in Contextual Embeddings for Sign Language Generation

Carla Viegas1,2    Mert Inan3    Lorna Quandt4    Malihe Alikhani 3
1Stella AI, Pittsburgh, USA
2Language Technologies Institute, Carnegie Mellon University, Pittsburgh, USA
3 Computer Science Department, School of Computing and Information,
University of Pittsburgh, Pittsburgh, USA
4 Educational Neuroscience Program, Gallaudet University, Washington, D.C, USA
Abstract

State-of-the-art sign language generation frameworks lack expressivity and naturalness which is the result of only focusing manual signs, neglecting the affective, grammatical and semantic functions of facial expressions. The purpose of this work is to augment semantic representation of sign language through grounding facial expressions. We study the effect of modeling the relationship between text, gloss, and facial expressions on the performance of the sign generation systems. In particular, we propose a Dual Encoder Transformer able to generate manual signs as well as facial expressions by capturing the similarities and differences found in text and sign gloss annotation. We take into consideration the role of facial muscle activity to express intensities of manual signs by being the first to employ facial action units in sign language generation. We perform a series of experiments showing that our proposed model improves the quality of automatically generated sign language.

Including Facial Expressions in Contextual Embeddings for Sign Language Generation


Carla Viegas1,2and Mert Inan3and Lorna Quandt4and Malihe Alikhani 3 1Stella AI, Pittsburgh, USA 2Language Technologies Institute, Carnegie Mellon University, Pittsburgh, USA 3 Computer Science Department, School of Computing and Information, University of Pittsburgh, Pittsburgh, USA 4 Educational Neuroscience Program, Gallaudet University, Washington, D.C, USA


1 Introduction

Communication between the Deaf and Hard of Hearing (DHH) people and hearing non-signing people may be facilitated by the emerging language technologies. DHH individuals are medically underserved worldwide McKee et al. (2020); Masuku et al. (2021) due to the lack of doctors who can understand and use sign language. Also, educational resources that are available in sign language are limited especially in STEM fields Boyce et al. (2021); Lynn et al. (2020). Although the Americans with Disabilities Act United States Department of Justice (2010) requires government services, public accommodations, and commercial facilities to communicate effectively with DHH individuals, the reality is far from ideal. Sign language interpreters are not always available and communicating through text is not always feasible as written languages are completely different from signed languages.

Refer to caption
Figure 1: Sign Language uses multiple modalities, such as hands, body, and facial expressions to convey semantic information. Although gloss annotation is often used to transcribe sign language, the above examples show that meaning encoded through facial expressions are not captured. In addition, the translation from text (blue) to gloss (red) is lossy even though sign languages have the capability to express the complete meaning from text. The lower example shows lowered brows and a wrinkled nose to add the meaning of kräftiger(heavy) (present in text) to the rain sign.

In contrast to Sign Language Recognition (SLR) which has been studied for several decades Rastgoo et al. (2021) in the computer vision community Yin et al. (2021), Sign Language Generation (SLG) is a more recent and less explored research topic  Quandt et al. (2021); Cox et al. (2002); Glauert et al. (2006).

Missing a rich grounded semantic representation, the existing SLG frameworks are far from generating understandable and natural sign language. Sign languages use spatiotemporal modalities and encode semantic information in manual signs and also in facial expressions. A major focus in SLG has been put on manual signs, neglecting the affective, grammatical, and semantic roles of facial expressions. In this work, we bring insights from computational linguistics to study the role of facial expressions in automated SLG. Apart from using facial landmarks encoding the contours of the face, eyes, nose, and mouth, we are the first to explore the use of facial Action Units (AUs) to learn semantic spaces or representations for sign language generation.

In addition, with insights from multimodal Transformer architecture design, we present a novel model, the Dual Encoder Transformer for SLG, which takes as input spoken text and glosses, computes the correlation between both inputs, and generates skeleton poses with facial landmarks and facial AUs. Previous work used either gloss or text to generate sign language or used text-to-gloss (T2G) prediction as an intermediary step Saunders et al. (2020). Our model architecture, on the other hand, allows us to capture information otherwise lost when using gloss only, and captures differences between text and gloss, which is especially useful for highlighting adjectives otherwise lost in gloss annotation. We perform several experiments using the PHOENIX14-T weather forecast dataset and show that our model performs better than baseline models using only gloss or text.

In summary, our main contributions are the following:

  • Novel Dual Encoder Transformer for SLG which captures information from text and gloss, as well as their relationship to generate continuous 3D sign pose sequences, facial landmarks, and facial action units.

  • Use of facial action units to ground semantic representation in sign language.

2 Background and Related Work

More than 70 million Deaf and Hard of Hearing worldwide use one of 300 existing sign languages as their primary language Kozik (2020). In this section, we explain the linguistic characteristics of sign languages, the importance of facial expressions to convey meaning, and elaborate on prior work in SLG.

2.1 Sign Language Linguistics

Sign languages are spatiotemporal languages and are articulated by using the hands, face, and other parts of the body, which need to be visible. In contrast to spoken languages which are oral-aural languages, sign languages are articulated in front of the top half of the body and around the head. No universal method such as the International Phonetic Alphabet (IPA) exists to capture the complexity of signs. Gloss annotation is often used to represent the meaning of signs in written form. Glosses do not provide any information about the execution of the sign, only about its meaning. Even more, as glosses use written languages rather than the sign language, they are a mere approximation of the sign’s meaning, representing only one possible transcription. For that reason, glosses do not always represent the full meaning of signs as shown in Figure 1.

Every sign can be broken into four manual characteristics: shape, location, movement, and orientation. Non-manual components such as mouth movements (mouthing), facial expressions, and body movements are other aspects of sign language phonology. In contrast to spoken languages, signing occurs simultaneously while vowels and consonants occur sequentially. Although the vocabulary size of ASL in dictionaries is around 15,000 Spread the Sign (2017) compared to approximately 170,000 in spoken English, the simultaneity of phonological components allows for a wide range of signs to describe slight differences of the same gloss.

While in English various words describe largeness (big, large, huge, humongous, etc.) in ASL, there is one main sign for “large”: big. However, through modifications of facial expressions, mouthing, and the size of the sign, different levels of largeness can be expressed just as in a spoken language Grushkin (2017). To communicate spoken concepts without a corresponding sign fingerspelling—a manual alphabet—is sometimes used. Baker et al. (2016)

NOUN VERB ADV ADJ
gloss 20927 6407 17718 648
TEXT 25952 7638 24755 5628
Table 1: Occurrence of different Part-of-Speech (POS) in the sign gloss annotation and the German transcripts computed with Spacy Honnibal and Montani (2017). Although gloss annotations show fewer samples for all POS, the difference in the occurrence of adjectives is statistically significant with p<0.05p<0.05.

2.2 Grammatical Facial Expressions

Facial expressions are grammatical components of sign languages that encode semantic representations, which, when excluded leads to loss of meaning. Facial expressions in particular have an important role in distinguishing different types of sentences such as WH-questions, Yes/No questions, doubt, negations, affirmatives, conditional clauses, focus and relative clauses da Silva et al. (2020). The following example shows how the same gloss order can present a question or an affirmation Baker et al. (2016):

Example 1

Indopakistani Sign Language
a) father car exist.
“(My) father has a car.”
b) father car exist?
“Does (your/his) father have a car.”

In this example, what makes sentence b) a question are raised eyebrows and a forward and/or downward movement of the head/chin in parallel to the manual signs.

Refer to caption
Figure 2: Examples from different facial Action Units (AUs) Friesen and Ekman (1978) from the lower face relevant to the generation of mouthings in sign languages. AUs can occur with different intensity values between 0 and 5. AUs have been used in psychology and in affective computing to understand emotions expressed through facial expressions. Image from De la Torre and Cohn (2011).

In addition, facial expressions can differentiate the meaning of a sign assuming the role of a quantifier. Figure 1 shows different signs for the same gloss, regen (rain). We can observe from the text transcript (in blue) that the news anchor says “rain” in the upper example but “heavy rain” in the lower. This example shows how gloss annotations are not perfect transcriptions of sign languages as they only convey the meaning of manual aspect of the signs. Information conveyed through facial expressions to show intensities are not represented in gloss annotation. To view the loss of information that occurs in gloss annotation we used Spacy Honnibal and Montani (2017) to compute the Part-of-Speech (POS) annotation for text and gloss. In Table 1 the occurrence of nouns, verbs, adverbs, and adjectives are shown for text and gloss over the entire dataset. We can see that although gloss annotations have lower occurrence for all POS, the difference is statistically significant for adjectives with p<0.05p<0.05. To calculate this significance, we performed hypothesis testing with two proportions by computing the Z score. We used t-tests to determine statistical significance of our model’s performance.

Refer to caption
Figure 3: Our proposed model architecture, the Dual Encoder Transformer for Sign Language Generation. Our architecture is characterized by using two encoders, one for text and one for gloss annotation. The use of two encoders allows to multiply the outputs of both emphasizing the differences and similarities. In addition we to using skeleton poses and facial landmarks, we include facial action units Friesen and Ekman (1978).

2.3 Sign Language Generation

Several advances in generating sign poses from text have been recently achieved in SLG, however there is limited work that considers the loss of semantic information when using gloss to generate poses and aligned facial expressions. Previous work has generated poses by translating text-to-gloss (T2G) and then gloss-to-pose (G2S) or by using either text or gloss as input Stoll et al. (2020); Saunders et al. (2020). We propose a Dual Encoder Transformer for SLG which trains individual encoders for text and gloss, and combines the encoder’s output to capture similarities and differences.

In addition, the majority of previous work on SLG has focused mainly on manual signs Stoll et al. (2020); Saunders et al. (2020); Zelinka and Kanis (2020); Saunders et al. (2021b)Saunders et al. (2021a) are the first to generate facial expressions and mouthing together with hand poses. The representation used for the non-manual channels is the same as for the hand gestures, namely coordinates of facial landmarks. In this work we explore the use of facial Action Units (AUs) (see Figure 2) which represent intensities of facial muscle movements Friesen and Ekman (1978). Although AUs have been primarily used in tasks related to emotion recognition Viegas et al. (2018), recent works have shown that AUs help detect WH-questions, Y/N questions, and other types of sentences in Brazilian Sign Language da Silva et al. (2020).

3 Sign Language Dataset

In this work we use the publicly available PHOENIX14T dataset Camgoz et al. (2018), frequently used as a benchmark dataset for SLR and SLG tasks. The dataset comprises a collection of weather forecast videos in German Sign Language (DGS), segmented into sentences and accompanied by German transcripts from the news anchor and sign-gloss annotations. PHOENIX14T contains videos of 9 different signers with 1066 different sign glosses and 2887 different German words. The video resolution is 210 by 260 pixels per frame and 30 frames per second. The dataset is partitioned into training, validation, and test set with respectively 7,096, 519, and 642 sentences.

4 Methods: Dual Encoder Transformer for Sign Language Generation

In this section, we present our proposed model, the Dual Encoder Transformer for Sign Language Generation. Given the loss of information that occurs when translating from text-to-gloss, our novel architecture takes into account the information from text and gloss as well as their similarities and differences to generate sign language in the form of skeleton poses and facial landmarks shown in Figure 3. For that purpose, we learn the conditional probability p=(Y|X,Z)p=(Y|X,Z) of producing a sequence of signs Y=(y1,,yT)Y=(y_{1},\ldots,y_{T}) with TT frames, given the text of a spoken language sentence XT=(x1,,xN)X_{T}=(x_{1},\ldots,x_{N}) with NN words and the corresponding glosses Z=(z1,,zU)Z=(z_{1},\ldots,z_{U}) with UU glosses.

Our work is inspired by the Progressive Transformer Saunders et al. (2020) which allows translation from a symbolic representation (words or glosses) to a continuous domain (joint and face landmark coordinates), by employing positional encoding to permit the processing of inputs with varied lengths. In contrast to the Progressive Transformer which uses one encoder to use either text or glosses to generate skeleton poses, we employ two encoders, one for text and one for glosses, to capture information from both sources, and create a combined representation from the encoder outputs to represent correlations between text and glosses. In the following we will describe the different components of the dual encoder transformer.

4.1 Embeddings

As our input sources are words, we need to convert them into numerical representations. Similar to transformers used for text-to-text translations, we use word embeddings based on the vocabulary present in the training set. As we are using two encoders to represent similarities and differences between text and glosses we use one word embedding based on the vocabulary of the text and one using the vocabulary of the glosses. We also experiment by using the text word embedding for both encoders. Given that our target is a sequence of skeleton joint coordinates, facial landmark coordinates, and continuous values of facial AUs with varying length we use counter encoding Saunders et al. (2020). The counter cc varies between [0,1] with intervals proportional to the sequence length. It allows the generation of frames without an end token. The target joints are then defined as:

mt\displaystyle m_{t} =[yt,ct] with\displaystyle=[y_{t},c_{t}]\textrm{ with }
yt\displaystyle y_{t} =[yhands+body,yface,yfacialAUs]\displaystyle=[y_{hands+body},y_{face},y_{facialAUs}]

The target joints mtm_{t} are then passed to a continuous embedding which is a linear layer.

4.2 Dual Encoders

We use two encoders, one for text and one for gloss annotations. Both encoders have the same architecture. They are composed by LL layers each with one Multi-head Attention (MHA) and a feed-forward layer. Residual connections He et al. (2016) around each of the two sublayers with subsequent layer normalization Ba et al. (2016). MHA uses multiple projections of scaled dot-products which permits the model to associate each word of the input with each other. The scaled dot-product attention outputs a vector of values, VV, which is weighted by queries, QQ, keys, KK, and dimensionality, dkd_{k}:

Attention(Q,K,V)=softmax(QKTdk)Attention(Q,K,V)=softmax(\frac{QK^{T}}{\sqrt{d_{k}}}) (1)

Different self-attention heads are used in MHA, which allows to generate parallel mappings of the QQ, VV, and KK with different learnt parameters.

The outputs of MHA are then fed into a non-linear feed-forward projection. In our case, where we employ two different encoders, their outputs can be formulated as:

Hn=Etext(w^n,w^1:N)Hu=Egloss(w^u,w^1:U)\begin{split}H_{n}=E_{text}(\hat{w}_{n},\hat{w}_{1:N})\\ H_{u}=E_{gloss}(\hat{w}_{u},\hat{w}_{1:U})\end{split} (2)

with hnh_{n} being the contextual representation of the source sequence, NN the number of words, and UU the number of glosses in the source sequence.

As we want to not only use the information encoded in text and gloss, but also their relationship, we combine the output of both encoders with a Hadamard multiplication. As the NUN\neq U, we stack hnh_{n} vertically for UU times and stack huh_{u} vertically for NN times in order to have two matrices with the same dimensions. Then we multiply both matrices with the Hadamard multiplication. Hadamard multiplication is a concatenation of every element in two matrices, where ai,ja_{i,j} and bi,jb_{i,j} are multiplied together to get ai,jbi,ja_{i,j}b_{i,j}. This represents concatenating the output vectors from the text encoder with the output of the vectors from the gloss encoder.

Htext,gloss=[Hn0Hn1HnU][Hu0Hu1HuN]H_{text,gloss}=\begin{bmatrix}H_{n0}\\ H_{n1}\\ \vdots\\ H_{nU}\end{bmatrix}\odot\begin{bmatrix}H_{u0}\\ H_{u1}\\ \vdots\\ H_{uN}\end{bmatrix} (3)

4.3 Decoder

Our decoder is based on the progressive transformer decoder (DPT), an auto-regressive model that produces continuous sequences of sign pose and the previously described counter value (Saunders et al., 2020). In addition to producing sign poses and facial landmarks, our decoder also produces 17 facial AUs. The counter-concatenated joint embeddings which include manual and facial features (facial landmarks and AUs), j^u\hat{j}_{u} , are used to represent the sign pose of each frame. Firstly, an initial MHA sub-layer is applied to the joint embeddings, similar to the encoder but with an extra masking operation. The masking of future frames is necessary to prevent the model from attending to future time steps. A further MHA mechanism is then used to map the sym- bolic representations from the encoder to the continuous domain of the decoder. A final feed forward sub-layer follows, with each sub-layer followed by a residual connection and layer normalisation as in the encoder. The output of the progressive decoder can be formulated as:

[y^u,c^u]=D(j^1:u1,h1:T)[\hat{y}_{u},\hat{c}_{u}]=D(\hat{j}_{1:u-1},h_{1:T}) (4)

where y^u\hat{y}_{u} corresponds to the 3D joint positions, facial landmarks, and AUs, representing the produced sign pose of frame uu and c^u\hat{c}_{u} is the respective counter value. The decoder learns to generate one frame at a time until the predicted counter value, c^u\hat{c}_{u}, reaches 1. The model is trained using the mean squared error (MSE) loss between the predicted sequence, y^1:U\hat{y}_{1:U} , and the ground truth, y1:Uy^{\ast}_{1:U} :

LMSE=1U(y1:Uy^1:U)2L_{MSE}=\frac{1}{U}(y^{\ast}_{1:U}-\hat{y}_{1:U})^{2} (5)
Components Dev Set Test Set
Bleu1 Bleu2 Bleu3 Bleu4 ROUGE Bleu1 Bleu2 Bleu3 Bleu4 ROUGE
Manual
30.15 20.58 15.41 12.22 30.41 27.76 18.86 14.11 11.32 27.44
Manual and Facial
29.46 20.30 15.31 12.10 29.25 26.75 17.88 13.29 10.61 26.54
Table 2: Translation results of the SLT model Camgoz et al. (2020) used for backtranslation when trained and evaluated with ground truth hand and body skeleton joints (manual) and facial landmarks and AUs (facial).
Model Dev Set Test Set
Bleu1 Bleu2 Bleu3 Bleu4 ROUGE Bleu1 Bleu2 Bleu3 Bleu4 ROUGE
G2S 24.51 15.71 11.19 8.70 24.84 23.26 14.54 10.21 7.84 22.89
T2S 22.90 14.55 10.42 8.14 23.42 22.14 13.88 9.85 7.56 22.50
TG2S (Ours) 24.60 16.20 11.68 8.97 24.82 22.97 14.71 10.59 8.19 23.45
Table 3: Back translation results obtained from the generative models when using only manual features. Our proposed model has the highest scores in almost all metrics compared to the models using only gloss or text.
Model Dev Set Test Set
Bleu1 Bleu2 Bleu3 Bleu4 ROUGE Bleu1 Bleu2 Bleu3 Bleu4 ROUGE
G2S 16.11 8.77 5.97 4.49 16.19 16.29 9.20 6.37 4.93 16.73
T2S 15.65 8.35 5.76 4.44 15.65 14.12 7.76 5.53 4.39 14.82
TG2S 17.25 10.17 7.04 5.32 17.85 17.18 10.39 7.39 5.76 17.64
Table 4: Back translation results obtained from the generative models when using manual features and facial landmarks and AUs. Our proposed model has the highest scores in all metrics compared to the models using only gloss or text.

5 Computational Experiments

5.1 Features

We extract three different types of features from the PHOENIX14T dataset: skeleton joint coordinates, facial landmark coordinates, and facial action unit intensities. We use OpenPose Cao et al. (2019) to extract skeleton poses from each frame and use for our experiments the coordinates of 50 joints which represent the upper body, arms, and hands, which we will start referring to as “manual features”. We also use OpenFace Baltrusaitis et al. (2018) to extract 68 facial landmarks as well as 17 facial action units (AUs) shown in Figure 2 to describe “facial features”.

5.2 Baseline Models

We will compare the performance of our proposed model (TG2S) with two Progressive Transformers Saunders et al. (2020), one using gloss only to produce sign poses (G2S), and one that uses text only (T2S). We train each model only with manual features and also with the combination of manual and facial features through concatenation.

5.3 Evaluation Methods

In order to automatically evaluate the performance of our model and the baseline models, we use back translation suggested by Saunders et al. (2020). For that purpose we use the Sign Language Transformer (SLT) Camgoz et al. (2020) which translates sign poses into text and computes BLEU and ROUGE scores between the translated text and the original text. As the original SLT was designed to receive video frames as input, we modified the architecture to enable the processing of skeleton poses and facial features.

Refer to caption
Figure 4: Comparison of the ground truth and the generated poses with our proposed dual encoder model for the gloss annotations cloud and rain. The upper example shows that the predictions captured the correct hand shape, orientation, and movement of the sign cloud. In the lower example it is visible that the predictions captured the repeating hand movement meaning rain. Although at first glance the hand orientation seems not correct, it is a slight variation which still is correct.
Refer to caption
Figure 5: Examples in which our model failed to generate the correct phonology of signs. Example 1 depicts inaccuracies in hand shape, orientation, and movement. Example 2 shows the difficulty of the model to capture pointing hand shapes.

6 Results

6.1 Quantitative Results

Table 2 shows how well the SLT model performs the translation from ground truth sign poses to text when trained and evaluated with the PHOENIX14T dataset. The results show the highest BLEU scores are achieved when training the SLT model only with skeleton joints from hands and upper body, presenting a BLEU-4 score of 11.32 for test set. When facial AUs are added to the hands, body, and face features, the difference from using manual data only is slightly lower, being BLEU-4 of 10.61.

In Table 3 the results of using hands and body joint skeleton as sole input to the baseline models and our proposed model are shown. We can see that our proposed model TG2S shows the highest BLEU-4 scores of 8.19 in test set, compared to 7.84 for G2S and 7.56 for T2S.

Table 4 presents the results of including facial landmarks as well as facial AUs with body and hands skeleton joints as input. Also here we can see that our proposed model outperforms the baseline models showing BLEU-4 score of 5.76 in test set. G2S obtained BLUE-4 score of 6.37 and T2S 5.53.

We see in Tables 3 and  4 that G2S obtained higher scores than T2S. Given that gloss annotations fail to encode the richness of meaning in signs, it appears the smaller vocabulary helps the model achieve higher scores by neglecting information otherwise described in text. Our proposed model is able to obtain better results than G2S by making a compromise of using information from gloss, text, and their similarities and differences. We also can see in both tables that the inclusion of facial information reduces the overall scores. We believe that this might be the case due to the diverse range of facial expressions possible. We cannot directly compare the results of Table 3 and  4 as two different SLT models were used to compute the BLEU scores.

6.2 Qualitative Results

Figure 4 shows the visual quality of our models prediction when using manual and facial information. Both examples show that the predictions captured the hand shape, orientation and movement from ground truth. In the bottom example for rain, the predictions were even able to capture the repetitive hand movement symbolizing falling rain. What can also be noted is that the ground truth is not perfect. In both examples unnatural finger and head postures can be seen. In addition, ground truth is not displaying movements of the eyebrows and mouth in the expected intensities.

Figure 5 shows situations in which the predictions failed to represent the correct phonology of signs. In the first example we see that hand shape, orientation, and position are not correct. The predictions of our models also fail to capture pointing hand shapes as shown in example 2.

7 Discussion and Conclusion

In this work, for the first time, we attempt to augment contextual embeddings for sign language by learning a joint meaning representation that includes fine-grained facial expressions. Our results show that the proposed semantic representation is richer and linguistically grounded.

Although our proposed model helped bridge the loss of information by taking into account text, gloss, and their similarities and differences, there are still several challenges to be tackled by a multidisciplinary scientific community.

Complex hand shapes with pointing fingers are very challenging to generate. The first step to improve the generation of the fingers is in improving methods to recognize finger movements more accurately. Similarly, we need tools that are more robust in detecting facial expressions even in situations of occlusion. We also realize that SLG models are overfitting specific sign languages instead of learning a generalized representations of signs.

We chose to work with a German sign language since that is the only dataset with gloss annotation that could help us study our hypotheses. The How2Sign dataset Duarte et al. (2021) is a feasible dataset for ASL, but it does not allow any model to extract facial landmarks, facial action units or facial expression from the original video frames since the faces are blurred. In the future, we hope to see new datasets with better and more diverse annotations for different sign languages that would allow the design of natural and usable sign language generation system.

Acknowledgements

We want to thank Stella AI LLC for supporting Carla Viegas with computational resources and funding. This project was partly supported by the University of Pittsburgh Momentum fund for research towards reducing language obstacles that Deaf students face when developing scientific competencies. We also acknowledge the Center for Research Computing at the University of Pittsburgh for providing part of the required computational resources. The author affiliated with Gallaudet University was partly supported by NSF Award IIS-2118742.

References

  • Ba et al. (2016) Jimmy Lei Ba, Jamie Ryan Kiros, and Geoffrey E Hinton. 2016. Layer normalization. arXiv preprint arXiv:1607.06450.
  • Baker et al. (2016) Anne Baker, Beppie van den Bogaerde, Roland Pfau, and Trude Schermer. 2016. The linguistics of sign languages: An introduction. John Benjamins Publishing Company.
  • Baltrusaitis et al. (2018) Tadas Baltrusaitis, Amir Zadeh, Yao Chong Lim, and Louis-Philippe Morency. 2018. Openface 2.0: Facial behavior analysis toolkit. In 2018 13th IEEE international conference on automatic face & gesture recognition (FG 2018), pages 59–66. IEEE.
  • Boyce et al. (2021) Richard Boyce, Malihe Alikhani, Sheila Pratt, David Boone, and Kenneth De Haan. 2021. Reducing language obstacles that deaf students face when developing scientific competencies. In Pitt Momentum Fund 2021.
  • Camgoz et al. (2018) Necati Cihan Camgoz, Simon Hadfield, Oscar Koller, Hermann Ney, and Richard Bowden. 2018. Neural sign language translation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 7784–7793.
  • Camgoz et al. (2020) Necati Cihan Camgoz, Oscar Koller, Simon Hadfield, and Richard Bowden. 2020. Sign language transformers: Joint end-to-end sign language recognition and translation. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 10023–10033.
  • Cao et al. (2019) Z. Cao, G. Hidalgo Martinez, T. Simon, S. Wei, and Y. A. Sheikh. 2019. Openpose: Realtime multi-person 2d pose estimation using part affinity fields. IEEE Transactions on Pattern Analysis and Machine Intelligence.
  • Cox et al. (2002) Stephen Cox, Michael Lincoln, Judy Tryggvason, Melanie Nakisa, Mark Wells, Marcus Tutt, and Sanja Abbott. 2002. Tessa, a system to aid communication with deaf people. In Proceedings of the fifth international ACM conference on Assistive technologies, pages 205–212.
  • da Silva et al. (2020) Emely Pujólli da Silva, Paula Dornhofer Paro Costa, Kate Mamhy Oliveira Kumada, José Mario De Martino, and Gabriela Araújo Florentino. 2020. Recognition of affective and grammatical facial expressions: a study for brazilian sign language. In European Conference on Computer Vision, pages 218–236. Springer.
  • De la Torre and Cohn (2011) Fernando De la Torre and Jeffrey F Cohn. 2011. Facial expression analysis. Visual analysis of humans, pages 377–409.
  • Duarte et al. (2021) Amanda Duarte, Shruti Palaskar, Lucas Ventura, Deepti Ghadiyaram, Kenneth DeHaan, Florian Metze, Jordi Torres, and Xavier Giro-i Nieto. 2021. How2sign: a large-scale multimodal dataset for continuous american sign language. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 2735–2744.
  • Friesen and Ekman (1978) E Friesen and Paul Ekman. 1978. Facial action coding system: a technique for the measurement of facial movement. Palo Alto, 3(2):5.
  • Glauert et al. (2006) JRW Glauert, Ralph Elliott, SJ Cox, Judy Tryggvason, and Mary Sheard. 2006. Vanessa–a system for communication between deaf and hearing people. Technology and Disability, 18(4):207–216.
  • Grushkin (2017) Don Grushkin. 2017. How large is the vocabulary of asl.
  • He et al. (2016) Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. 2016. Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 770–778.
  • Honnibal and Montani (2017) Matthew Honnibal and Ines Montani. 2017. spaCy 2: Natural language understanding with Bloom embeddings, convolutional neural networks and incremental parsing. To appear.
  • Kozik (2020) Karolina Kozik. 2020. Without sign language, deaf people are not equal.
  • Lynn et al. (2020) Matthew A Lynn, David C Templeton, Annemarie D Ross, Austin U Gehret, Morgan Bida, Timothy J Sanger, and Todd Pagano. 2020. Successes and challenges in teaching chemistry to deaf and hard-of-hearing students in the time of covid-19. Journal of Chemical Education, 97(9):3322–3326.
  • Masuku et al. (2021) Khetsiwe P Masuku, Nomfundo Moroe, and Danielle van der Merwe. 2021. ‘the world is not only for hearing people–it’s for all people’: The experiences of women who are deaf or hard of hearing in accessing healthcare services in johannesburg, south africa. African Journal of Disability, 10.
  • McKee et al. (2020) Michael McKee, Christa Moran, and Philip Zazove. 2020. Overcoming additional barriers to care for deaf and hard of hearing patients during covid-19. JAMA Otolaryngology–Head & Neck Surgery, 146(9):781–782.
  • Quandt et al. (2021) Lorna C Quandt, Athena Willis, Melody Schwenk, Kaitlyn Weeks, and Ruthie Ferster. 2021. Attitudes toward signing human avatars vary depending on hearing status, age of signed language exposure, and avatar type.
  • Rastgoo et al. (2021) Razieh Rastgoo, Kourosh Kiani, and Sergio Escalera. 2021. Sign language recognition: A deep survey. Expert Systems with Applications, 164:113794.
  • Saunders et al. (2020) Ben Saunders, Necati Cihan Camgoz, and Richard Bowden. 2020. Progressive transformers for end-to-end sign language production. In European Conference on Computer Vision, pages 687–705. Springer.
  • Saunders et al. (2021a) Ben Saunders, Necati Cihan Camgoz, and Richard Bowden. 2021a. Continuous 3d multi-channel sign language production via progressive transformers and mixture density networks. International Journal of Computer Vision, pages 1–23.
  • Saunders et al. (2021b) Ben Saunders, Necati Cihan Camgoz, and Richard Bowden. 2021b. Mixed signals: Sign language production via a mixture of motion primitives.
  • Spread the Sign (2017) Spread the Sign. 2017. About us.
  • Stoll et al. (2020) Stephanie Stoll, Necati Cihan Camgoz, Simon Hadfield, and Richard Bowden. 2020. Text2sign: towards sign language production using neural machine translation and generative adversarial networks. International Journal of Computer Vision, 128(4):891–908.
  • United States Department of Justice (2010) United States Department of Justice. 2010. Ada requirements: Effective communication.
  • Viegas et al. (2018) Carla Viegas, Shing-Hon Lau, Roy Maxion, and Alexander Hauptmann. 2018. Towards independent stress detection: A dependent model using facial action units. In 2018 International Conference on Content-Based Multimedia Indexing (CBMI), pages 1–6. IEEE.
  • Yin et al. (2021) Kayo Yin, Amit Moryossef, Julie Hochgesang, Yoav Goldberg, and Malihe Alikhani. 2021. Including signed languages in natural language processing. arXiv preprint arXiv:2105.05222.
  • Zelinka and Kanis (2020) Jan Zelinka and Jakub Kanis. 2020. Neural sign language synthesis: Words are our glosses. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, pages 3395–3403.