This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Multi-modal Automated Speech Scoring using Attention Fusion

Abstract

In this study, we propose a novel multi-modal end-to-end neural approach for automated assessment of non-native English speakers’ spontaneous speech using attention fusion. The pipeline employs Bi-directional Recurrent Convolutional Neural Networks and Bi-directional Long Short-Term Memory Neural Networks to encode acoustic and lexical cues from spectrograms and transcriptions, respectively. Attention fusion is performed on these learned predictive features to learn complex interactions between different modalities before final scoring. We compare our model with strong baselines and find combined attention to both lexical and acoustic cues significantly improves the overall performance of the system. Further, we present a qualitative and quantitative analysis of our model.

Index Terms—  automated speech scoring, spontaneous speech, end-to-end, multi-modal, attention fusion

1 Introduction

Automated Scoring (AS), in general, refers to the act of using computers to convert examinee’s performance on standardized tests to some performance indicators. The questions types can range from multiple choice questions to written and spoken language assessment. AS systems are expected to pick up relevant signals from examinee’s response and transform them to metrics which accurately infer their ability [1]. The task of speech scoring, specifically, is the task of standardized assessment of speaking proficiency for speakers of a language. Such assessments can then inform important decisions like the hiring of candidates for a company or admission to schools, judging academic proficiency levels, etc. The history of AS systems is quite old with its humble beginnings in 1964 when Page and colleagues [2, 3] started scoring essay on punch cards.

The earlier systems judged the performance of L2 speakers by making them recite a written text [4]. These recitals were compared with the speech of native speakers, and scores were produced accordingly. They gradually became more complex and started including handcrafted features for measuring features like pronunciation, fluency, stress, intonation and content [5]. Although this worked in practice, yet this type of testing was not able to judge complex features like opinion formation, argument depth, structure of the prose, etc. With advances in machine learning and ASR, the type of testing changed from just recital to more complex forms involving open-ended questions [6]. Lately, studies have shown end-to-end systems performing better when compared to feature-based systems for this type of testing. In addition, such systems alleviate the development effort needed for scoring these responses [6, 7, 8, 9]. While these black box systems allow automated extraction of meaningful representations, very few studies, particularly in speech scoring, share insights of the predictions made. In this paper, we present a novel multi-modal end-to-end neural pipeline for automated oral proficiency assessment of L2 English speakers. We leverage a deep-learning-based ASR system for transcription and Bi-directional Recurrent Convolutional Neural Networks (BDRCNN) and Bi-directional Long Short-Term Memory Neural Networks (BDLSTM) to learn lexical and acoustic feature representations. We perform attention fusion [10] on these features to learn complex interactions among these features. For analysis, we leverage learned attention weights to understand the important parts of the response.

Refer to caption
Fig. 1: Multi-modal with Attention Fusion Speech Grading Architecture.
Table 1: Statistics of the dataset. P: Prompt Number, #R: Number of responses, D: Difficulty level, Sz: Average Response Size (Duration in seconds, Length in number of word tokens), and DS: Distribution of Scores (prefixes ‘L’ means Low and ‘H’ means High).
P #R D Sz DS
Duration Length A2 LB1 HB1 LB2 HB2
1 8008 B1 57.7 100 279 1575 6154 - -
2 8160 B1 58.7 108 546 3122 4492 - -
3 8169 B2 81.5 149 121 672 3535 3738 103
4 8137 C1 104.2 180 123 729 3573 3601 111
5 8100 C1 106.1 196 112 557 3061 4216 154
6 8158 B1 55.9 110 121 1045 6992 - -

2 Deep Learning based Scoring Models

2.1 Bi-directional Recurrent CNN (for audio)

An important aspect of scoring rubric is the delivery of a spoken response. It helps in measuring speech quality, pronunciation, fluency, stress, and intonation [5]. To capture this acoustic information of audio samples, we propose a Recurrent Convolutional Neural Network (RCNN) architecture. CNN is able to extract local features, whilst the RNN is able to summarise the long temporal information [11]. Such an architecture proves beneficial to us because of the nature of our data; which consists of long monologues (ranging from 60 to 120 seconds), and have long term interdependencies present in it.

We split the log-scaled mel spectrograms of the audio samples along the temporal dimension into frames of fixed size (discussed in Section 3.2). Each frame is passed to 5 sets of CNN consisting of 2 layers of 1D convolutions followed by max-pooling with the number of filters doubling after every set. Then, we apply global max pooling to each frame. The output vectors obtained from this network are then fed into the RNN component, which in our case is a BDLSTM allowing us to capture the sequential structure of audio, taking into account both past and future speech production.

2.2 Bi-directional Long Short-Term Memory (for text)

It is also essential that the content of the spoken response is relevant to the topic and appropriate. Given the nature of our data, we use response-based content scoring, where we compare the similarity of response with previously scored responses. This approach is popular in both automated scoring of essays [6, 12] as well as speech scoring [7] as it allows capturing variations in responses comprehensively.

To achieve this, we generate transcripts of spoken responses using a non-native ASR system. These transcripts are pre-processed, and words are mapped to their word embeddings using an embedding layer. This embedding layer is initialized with 300-D glove embeddings [13] trained on Wikipedia and is optimized during training. Similar to the acoustic model, we use BDLSTM network here to capture sequential the structure of words and learn the development of content at different score levels.

Table 2: Quadratic Kappa Score and Mean Squared Error across prompts. ( indicates threshold optimized models)
Prompt 1 Prompt 2 Prompt 3 Prompt 4 Prompt 5 Prompt 6
Model QWK MSE QWK MSE QWK MSE QWK MSE QWK MSE QWK MSE
BDRCNNAttn [A] 0.302 0.288 0.240 0.484 0.456 0.476 0.454 0.478 0.426 0.477 0.237 0.188
BDLSTMAttn [T] 0.473 0.240 0.305 0.443 0.490 0.455 0.564 0.388 0.549 0.389 0.402 0.148
MMAF 0.529 0.229 0.334 0.425 0.547 0.406 0.586 0.371 0.578 0.371 0.435 0.146
BDRCNNAttn [A] 0.412 0.318 0.279 0.511 0.455 0.518 0.471 0.520 0.452 0.499 0.236 0.208
BDLSTMAttn [T] 0.543 0.267 0.355 0.487 0.525 0.469 0.563 0.403 0.564 0.482 0.451 0.179
MMAF 0.550 0.269 0.373 0.465 0.539 0.473 0.606 0.372 0.597 0.445 0.480 0.160
Human-Human 0.676 0.186 0.564 0.328 0.766 0.241 0.785 0.213 0.823 0.177 0.687 0.099

2.3 Multi-modal with Attention Fusion (MMAF)

In the past few years, attention mechanism [14] has achieved state-of-the-art on various natural language processing tasks. It enables weighting contextual information learned during each time step, allowing the model to determine which states to pay attention to. In our study, we use attention fusion [10] to combine features from the text and audio modality for grading spoken responses (see Figure 1). Given bi-directional temporal hidden states from audio model (BDRCNN), hah^{a}, and text model (BDLSTM), hth^{t}, we combine these states, hm=[ha,ht]h^{m}=[h^{a},h^{t}] (where [][\cdot] denotes the concatenation of the state vectors) and compute context vector, cmc^{m} as:

et=htmwa;atm=exp(et)i=1Texp(ei);cm=t=1Tatmhtme_{t}=h_{t}^{m}w_{a};a_{t}^{m}=\frac{\exp(e_{t})}{\sum_{i=1}^{T}\exp(e_{i})};c^{m}=\sum_{t=1}^{T}a_{t}^{m}h_{t}^{m} (1)

Here, htmh_{t}^{m} is the multimodal representation of the lexical/acoustic cues at time tt and waw_{a} is the weight matrix for the attention layer. The multimodal attention importance scores for each time, atma_{t}^{m} is obtained by multiplying the representation htmh_{t}^{m} with the weight matrix, waw_{a}, followed by normalization to construct probability distribution across these cues. Finally, we calculate the context vector of cues, cmc^{m}, as a weighted summation over all time steps using multimodal attention importance scores as weights. This context vector is passed through a final dense layer before generating score for a input spoken response.

For uni-modal baselines, we applied attention weighting to individual BDRCNN and BDLSTM models (BDRCNNAttn [A] for audio and BDLSTMAttn [T] for text) and trained them separately for comparison.

3 Experiments

3.1 Data

In this study, we utilize data collected by Second Language Testing Inc. (SLTI) administrating Simulated Oral Proficiency Interview (SOPI) for L2 English speakers, majorly from the Philippines. Each examinee is presented with a form consisting of six prompts of varying difficulty levels on their computer screen, and their responses are recorded. These responses are then scored independently by two expert raters. A third expert rater resolves disagreements, if any. Both the prompts and rubrics are aligned with the guidelines of the Common European Framework of Reference (CEFR) [15]. To answer the questions, candidates require explanatory and argumentative abilities. Depending on their performance on these tasks, they are assessed to have English oral proficiency ranging from lowest grade, A2, to highest grade, C1 (with the maximum score obtainable being equal to the difficulty of the prompt). The examinees are mostly high school graduates of age 17 years and above. Most of the recorded audios have a size lesser than 120 seconds. Table 1 shares the other relevant statistics of the dataset. Dual-scoring with the intervention of third rater and robust rubric ensured that the scoring process is not biased towards irrelevant details like gender and features like text length. As a point of reference, the correlation between score and text length if 0.35 across all the prompts.

3.2 Experimental setup

For each prompt, we stratify and split the responses into train, validation and test with the ratio 70:10:2070:10:20. We train the model for every prompt and report the quadratic kappa score obtained on the test set. For training the models, first, we downsample the audio responses to 16kHz16kHz and generate their log-scaled mel spectrograms. The number of mel bands is 128128, the length of the FFT window is 20482048 samples, and the hop length is 512512 samples. These spectrograms are then normalized and zero-padded to maximum response length. Finally, we split the output along the temporal dimension into frames of size 128×128128\times 128.

We use a Deep Speech 2 [16] based ASR system with trigram language model for transcription of these non-native responses. The ASR system is trained on approximately 10001000 hours of audio sampled from CommonVoice [17] and LibriSpeech dataset [18] and further fine-tuned on approximately 2222 hours of transcribed non-native spoken responses from our dataset. This ASR system achieves a word error rate of 16.63%16.63\% on approximately 55 hours of non-native spoken responses.

We tokenize the transcripts using the spacy tokenizer [19] and lowercase them. Words not part of training data vocabulary are mapped to the unknown token and initialized with zero vector in the embedding layer. We treat the scoring of the responses as a regression problem. The CEFR scores (NN levels) associated with the responses are mapped to the range [0,N1][0,N-1] based on the increase in proficiency level. These scores are then normalized to the range of [0,1][0,1] for training. While testing, we rescale the model output to the original score range and measure the performance.

3.3 Evaluation metric

For evaluating the models, we use the Quadratic Weighted Kappa (QWK) score as our primary metric. The QWK measures the agreement between two graders. It normally ranges from 0 to 11 and can also be negative if there is lesser agreement than expected by chance. To calculate the QWK score, a weight matrix WW of size N×NN\times N is constructed, where NN represents the number of classes, using the formula as mentioned in Eq. 2.

Wi,j=(ij)2(N1)2W_{i,j}=\frac{(i-j)^{2}}{(N-1)^{2}} (2)

Next, we construct a confusion matrix OO of size N×NN\times N, where OijO_{ij} is equal to the number of speech responses that receive a grade ii by the human and a grade jj by the model.

Then we create the histogram matrix of expected grades EE by computing the outer product between the histogram vector of actual grades and the histogram vector of predicted grades, followed by a normalization which ensures that EE and OO have the same sum. Lastly, the QWK is obtained as in Eq. 3.

κ=1i,jWi,jOi,ji,jWi,jEi,j\kappa=1-\frac{\sum_{i,j}W_{i,j}O_{i,j}}{\sum_{i,j}W_{i,j}E_{i,j}} (3)

3.4 Training

We use the Adam optimization algorithm to minimize the mean squared error (MSE) loss over the training data (see Eq. 4).

MSE(ytrue,ypred)=1Ni=1N(ytrueiypredi)2MSE(y_{true},y_{pred})=\frac{1}{N}\sum_{i=1}^{N}(y_{true}^{i}-y_{pred}^{i})^{2} (4)

where NN is the number of audio responses for a given prompt, and ytrueiy_{true}^{i} and yprediy_{pred}^{i} are the normalized human grade and model predicted grade for response ii, respectively. To prevent overfitting of the model, we use dropout regularization. We train the model for a fixed number of epochs with early stopping on validation loss and select the model with the best QWK score on the validation set. We also optimize thresholds between each score for rounding off the predictions in order to maximize the QWK. This is done by defining the search space of thresholds between each possible score and optimizing for maximum QWK using the hyperopt package [20].

Table 3: Attention split across prompts. P: Prompt Number, TA: Text Attention and AA: Audio Attention
P TA AA
1 84.89% 15.11%
2 86.21% 13.79%
3 84.38% 15.62%
4 88.52% 11.48%
5 84.73% 15.27%
6 89.53% 10.47%
Table 4: Attention split across grades. P: Prompt Number, G: Grade, TA: Text Attention and AA: Audio Attention
P G TA AA
1 A2 81.63% 18.37%
Low B1 83.52% 16.48%
High B1 85.44% 14.56%
4 A2 87.20% 12.80%
Low B1 87.52% 12.48%
High B1 88.26% 11.74%
Low B2 88.97% 11.03%
High B2 89.69% 10.31%
Table 5: Quadratic Kappa Score with audios replaced by white noise (denoted by ) and by Text-to-Speech outputs of transcripts (denoted by ). P: Prompt Number, TO: Threshold Optimization
P Without TO With TO
1 0.446 0.487
4 0.399 0.409
1 0.420 0.451
4 0.388 0.388

4 Results and Discussion

4.1 Quantitative Evaluation

In Table 2, we compare MMAF model with two strong baselines: BDRCNNAttn [A] and BDLSTMAttn [T] models on the SOPI dataset. It shows that fusing both the modalities using attention gives the best scores across all the prompts. Overall, the improvement in average QWK score across prompts for MMAF when compared to BDLSTMAttn [T] and BDRCNNAttn [A] models is close to 8.2% and 42.6% respectively. Optimizing thresholds improves the scores by approximately 4.8% over text-only models and 36.5% over audio-only model. Since prompts (1, 2 and 6) have high class imbalance and QWK is not a reliable metric since it becomes too sensitive towards the under-represented classes [21, 22], we also track MSE for comparison. We observe the average MSE across prompts for MMAF model decreases by 5.5% and 18.3% when compared to BDLSTMAttn [T] and BDRCNNAttn [A] models respectively. Table 2 shows that the performance of all models for lower difficulty level prompts (prompt 1, 2 and 6) is poor compared to higher difficulty prompts (prompt 3, 4 and 5). The human-human agreement also shows similar performance. Table 5 shows the variation of attention scores across all the prompts. It is evident that attention weights prefer text over audio broadly in the ratio of 85:15. Due to the non-availability of any public dataset or model, we were not able to show the same results on those [7, 9].

Refer to caption
Fig. 2: Attention plots of frames taken from 2 samples of Prompt 4.

For further analysis, we select two exemplar prompts: one from the lower difficulty range (prompt 1) and one from the higher difficulty (prompt 4). Table 5 shows that audio attention weights partially increase for low scored responses. By checking the samples, we could infer that one of the reasons for this is due to the lower intelligibility of audios for those items. This, in turn, caused ASR word error rate to go up nudging the model to partially shift the attention to audio over the text modality.
We also investigate the effect of audio variation on our model by performing two sets of experiments.

1. White noise with intact text: Replacing audio inputs to the model with white noise, we observe that the QWK drops across both prompt 1 and 4, by an average of 23.8% and 22% without and with threshold optimization respectively (Table 5).

2. Template native voice with intact text: Next, we replace the audio inputs to the model with the Text-to-Speech outputs of the transcriptions. We use the eSpeak [23] speech synthesizer for this experiment (see Table 5 for results). We observe an average drop in QWK of 27.2% for the model without threshold optimization and of 27% for the model with threshold optimization, across both the prompts. This performance drop can possibly be explained due to the fact that the model was originally trained on speech responses from L2 speakers, and hence performed poorly when tested on L1 speech.

Curiously, the model scored the speech with the native accented templatized audio consistently lower than the speech with white noise audio part. In our analysis, this was majorly due to two reasons: text-to-speech system did not produce much sound variation and that the generated voice was closer to a native speaker but was tested on a model trained for L2 speakers.

4.2 Qualitative Evaluation

For checking the attention weights of the model on a more granular level, we manually analyzed some samples by plotting their attention weights. Figure 2 shows attention splits for one “Low B1” and one “Low B2” scored sample, both taken from prompt 4, which is a prompt based on the topic of smoking. We show two frames (or segmentations) for both the samples, one in each row. Each cell represents one frame of audio input and its corresponding transcription. The bar graph shows the respective audio and text attentions, normalized separately using min-max scaling across the entire response. We can observe that for filler words like “uh” and “um”, false starts and pauses, the model pays great attention to the audio. For content-rich fluent portions, the model chooses to pay significantly high attention to the transcripts. This is particularly visible in Figure 2 where we can see that model pays high attention to text for frames containing words like “no smoking”, which would be important to judge the response to the question.

5 Conclusion

In this paper, we investigated end-to-end deep-learning based models for the automated speech scoring task for L2 English speakers. We showed that multi-modal attention fusion works better than uni-modal networks. Results showed that transcript response of the speech is much more important as compared to audio and that it becomes progressively more important for higher scored samples. Future work would involve a feature-based analysis of all the responses and cross-question comparison of the responses on various difficulty levels. It would also be interesting to see how the ablation of network inputs and outputs fare with respect to features.

References

  • [1] Yaman Kumar, Swati Aggarwal, Debanjan Mahata, Rajiv Ratn Shah, Ponnurangam Kumaraguru, and Roger Zimmermann, “Get it scored using autosas—an automated system for scoring short answers,” in Proceedings of the AAAI Conference on Artificial Intelligence, 2019, vol. 33, pp. 9662–9669.
  • [2] Ellis B Page, “The imminence of… grading essays by computer,” The Phi Delta Kappan, vol. 47, no. 5, pp. 238–243, 1966.
  • [3] Ellis Batten Page, “Computer grading of student prose, using modern concepts and software,” The Journal of experimental education, vol. 62, no. 2, pp. 127–142, 1994.
  • [4] Silke M Witt and Steve J Young, “Phone-level pronunciation scoring and assessment for interactive language learning,” Speech communication, vol. 30, no. 2-3, pp. 95–108, 2000.
  • [5] Lei Chen, Klaus Zechner, Su-Youn Yoon, Keelan Evanini, Xinhao Wang, Anastassia Loukina, Jidong Tao, Lawrence Davis, Chong Min Lee, Min Ma, Robert Mundkowsky, Chi Lu, Chee Wee Leong, and Binod Gyawali, “Automated scoring of nonnative speech using the speechrater sm v. 5.0 engine,” ETS Research Report Series, vol. 2018, no. 1, pp. 1–31, 2018.
  • [6] Kaveh Taghipour and Hwee Tou Ng, “A neural approach to automated essay scoring,” in Proceedings of the 2016 conference on empirical methods in natural language processing, 2016, pp. 1882–1891.
  • [7] Lei Chen, Jidong Tao, Shabnam Ghaffarzadegan, and Yao Qian, “End-to-end neural network based automated speech scoring,” in 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, 2018, pp. 6234–6238.
  • [8] Dimitrios Alikaniotis, Helen Yannakoudakis, and Marek Rei, “Automatic text scoring using neural networks,” in Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Berlin, Germany, Aug. 2016, pp. 715–725, Association for Computational Linguistics.
  • [9] Y. Qian, P. Lange, K. Evanini, R. Pugh, R. Ubale, M. Mulholland, and X. Wang, “Neural approaches to automated speech scoring of monologue and dialogue responses,” in ICASSP 2019 - 2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 2019, pp. 8112–8116.
  • [10] Chiori Hori, Takaaki Hori, Teng-Yok Lee, Ziming Zhang, Bret Harsham, John R. Hershey, Tim K. Marks, and Kazuhiko Sumi, “Attention-based multimodal fusion for video description,” in The IEEE International Conference on Computer Vision (ICCV), Oct 2017.
  • [11] K. Choi, G. Fazekas, M. Sandler, and K. Cho, “Convolutional recurrent neural networks for music classification,” in 2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 2017, pp. 2392–2396.
  • [12] Yi Tay, Minh Phan, Luu Anh Tuan, and Siu Cheung Hui, “Skipflow: Incorporating neural coherence features for end-to-end automatic text scoring,” in AAAI Conference on Artificial Intelligence, 2018.
  • [13] Jeffrey Pennington, Richard Socher, and Christopher D. Manning, “Glove: Global vectors for word representation,” in Empirical Methods in Natural Language Processing (EMNLP), 2014, pp. 1532–1543.
  • [14] Dzmitry Bahdanau, Kyunghyun Cho, and Yoshua Bengio, “Neural machine translation by jointly learning to align and translate,” in 3rd International Conference on Learning Representations, ICLR 2015, San Diego, CA, USA, May 7-9, 2015, Conference Track Proceedings, 2015.
  • [15] Council of Europe. Council for Cultural Co-operation. Education Committee. Modern Languages Division, Common European Framework of Reference for Languages: learning, teaching, assessment, Cambridge University Press, 2001.
  • [16] Dario Amodei, Sundaram Ananthanarayanan, Rishita Anubhai, Jingliang Bai, Eric Battenberg, Carl Case, Jared Casper, Bryan Catanzaro, Jingdong Chen, Mike Chrzanowski, Adam Coates, Greg Diamos, Erich Elsen, Jesse Engel, Linxi Fan, Christopher Fougner, Awni Y. Hannun, Billy Jun, Tony Han, Patrick LeGresley, Xiangang Li, Libby Lin, Sharan Narang, Andrew Y. Ng, Sherjil Ozair, Ryan Prenger, Sheng Qian, Jonathan Raiman, Sanjeev Satheesh, David Seetapun, Shubho Sengupta, Chong Wang, Yi Wang, Zhiqian Wang, Bo Xiao, Yan Xie, Dani Yogatama, Jun Zhan, and Zhenyao Zhu, “Deep speech 2 : End-to-end speech recognition in english and mandarin,” in ICML, 2016, pp. 173–182.
  • [17] Rosana Ardila, Megan Branson, Kelly Davis, Michael Henretty, Michael Kohler, Josh Meyer, Reuben Morais, Lindsay Saunders, Francis M. Tyers, and Gregor Weber, “Common voice: A massively-multilingual speech corpus,” 2019.
  • [18] V. Panayotov, G. Chen, D. Povey, and S. Khudanpur, “Librispeech: An asr corpus based on public domain audio books,” in 2015 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 2015, pp. 5206–5210.
  • [19] Matthew Honnibal and Ines Montani, “spaCy 2: Natural language understanding with Bloom embeddings, convolutional neural networks and incremental parsing,” To appear, 2017.
  • [20] J. Bergstra, D. Yamins, and D. D. Cox, “Making a science of model search: Hyperparameter optimization in hundreds of dimensions for vision architectures,” in Proceedings of the 30th International Conference on International Conference on Machine Learning - Volume 28. 2013, ICML’13, p. I–115–I–123, JMLR.org.
  • [21] Alvan R Feinstein and Domenic V Cicchetti, “High agreement but low kappa: I. the problems of two paradoxes,” Journal of clinical epidemiology, vol. 43, no. 6, pp. 543–549, 1990.
  • [22] Domenic V Cicchetti and Alvan R Feinstein, “High agreement but low kappa: Ii. resolving the paradoxes,” Journal of clinical epidemiology, vol. 43, no. 6, pp. 551–558, 1990.
  • [23] “eSpeak: Speech synthesizer,” Mar 2014.