Multi-channel acoustic modeling using mixed bitrate opus compression
Abstract
Recent literature has shown that a learned front end with multi-channel audio input can outperform traditional beamforming algorithms for automatic speech recognition (ASR). In this paper, we present our study on multi-channel acoustic modeling using OPUS compression with different bitrates for the different channels. We analyze the degradation in word error rate (WER) as a function of the audio encoding bitrate and show that the WER degrades by 12.6% relative with 16kpbs as compared to uncompressed audio. We show that its always preferable to have a multi-channel audio input over a single channel audio input given limited bandwidth. Our results show that for the best WER, when one of the two channels can be encoded with a bitrate higher than 32kbps, its optimal to encode the other channel with the highest bitrate possible. For bitrates lower than that, its preferable to distribute the bitrate equally between the two channels. We further show that by training the acoustic model on mixed bitrate input, upto 50% of the degradation can be recovered using a single model.
Index Terms— multi-channel, opus, far-field speech recognition
1 Introduction
Multi-channel modeling has become a popular alternative to conventional beamforming in recent literature [1][2]. This method of front-end processing opens up avenues for processing audio not just from microphones from a single microphone array but from microphones from distinct devices. With the popularity of voice enabled devices in the consumer markets, this opens up an interesting area of research. Different devices from different manufacturers, however, may not adhere to the same encoding standards. In addition, the bandwidth available for audio upload at different times might be different. Motivated by these questions, we wanted to study how multi-channel acoustic models perform when different audio inputs have different bitrates.
Prior work in this area has focused at developing compression algorithms to minimally impact speech recognition systems [3] or at analyzing the quality of the opus codec as it pertains to voice quality [4] . In [5], the authors evaluate the performance of different audio codecs on the frequency spectrum. There is also prior work in ASR that examines mixed-bandwidth models that deal with data with different sampling rates [6]. There has been some related work in the emotion recognition domain; the authors in [7] and [8] show how different codecs and different bitrates affect emotion recognition accuracy. The authors in [9] analyze human emotion intelligibility as a function of the bitrates. The other related work is in the domain of multi-channel acoustic modeling where the traditional front-end processing algorithms like beamforming are learned within the acoustic modeling network. [10] and [2] demonstrate that a data driven beamformer out-performs a traditional beamforming approaches.
Our contribution in this paper is to study how the ASR model performance varies with the different bitrate encoding of the OPUS codec [11]. We chose the OPUS codec for our analysis because it is an open source, industry leading standard for audio coding for multiple applications [12]. The main benefit of OPUS over other codes is the hybrid approach, where it uses SILK for encoding information below 8khz and CELT for information above 8khz [11], thus providing high quality for both speech and audio/music components. We analyze the performance of both the multi-channel and single channel acoustic models as a function of the audio input bitrate. We demonstrate that by training with mixed-bitrates we can recover some of the performance loss. Motivated by the multiple device use case described above, we also study the performance of the system if the multiple audio input channels to the network are not encoded with the same bitrate. The closest work to this was in [13], where the authors show how the ASR WER varies with different codecs at a fixed bitrate encoding.
This paper is organized as follows. In Section 2, we describe our model architecture, training techniques and the training and evaluation data. In Section 3, we present our analysis of the ASR system performance with different bitrate audio inputs. We discuss our experiments on how to optimally distribute bandwidth between multiple channels in Section 4. In Section 5, we describe the mixed-bitrate model training focused at recovering from degradation introduced by low bitrate encoding. Finally in Section 6, we present analyze and conclude our work, and present future work.
2 System description
Our network architecture is similar to the Elastic Spatial Filtering system described in [10] and is shown in Figure 1. The network takes in the FFT from a multi-channel audio input. The first layer of the network is a block affine transform initialized by the coefficients of a super-directive beamformer that uses the spherically isotropic noise field [14][15]. This is followed by component that computes the power of each of the frequency components in each of the look directions. This is followed by an affine transform that performs elastic spatial filtering as described in [10]. This block is followed by another affine transform that is initialized by mel-filter bank (MFB) coefficients, followed by a ReLU non-linearity and log to emulate computing log filter-bank energies. The ReLU component is used to ensure positive value input to the log component. These components form the learned feature extraction layers. This is followed by LSTM layers that classify the input to senones. The LSTM weights are initialized using a uniform distribution.
For all the experiments in the paper, we use a 2-channel speech input and a 256-point FFT without the Nyquist frequency and direct components. These features are mean and variance normalized and are then used as input into the network. The block affine transform is initialized with 12 different look directions. The output of the MFB layer is 64-dimensional. For the classification layers we use a 5-layer LSTM with 768 cells in each layer. Our system uses 3183 senones. We use an effective frame size of 30ms; the FFT is computed from a 10ms frame, and the beamformer, elastic spatial filtering and MFB affine transform layer operate at a frame size of 10ms. After the MFB layer, the features for 3 frames are concatenated and the rest of the processing consumes a 30ms frame. Our training data consists of 620 hours of far-field data from our in-house dataset. The data is sampled at 16kHz. The array geometry used here is an equispaced six-channel microphone circular array with a diameter of approximately 72 milli-meters and one microphone at the center. For all our experiments, we use 2 microphones from the opposite sides of the microphone array. All the models are trained with the cross-entropy objective followed by the sMBR objective function [16]. Our test data consists of 33,000 far-field speech utterances, and our development set consists of 17,000 similar utterances.
3 Performance as a function of encoding bitrates
For this study, we train our baseline multi-channel model with uncompressed audio as input. The architecture and training data are described in Section 2. We use the development set to tune the acoustic scale for our system. For this specific study, our development set is also uncompressed audio. We use the OPUS encoder [17] to encode the uncompressed audio in our test set using different bitrates using the constant bitrate setting 111Our experiments didn’t show a significant difference between the variable bitrate and constant bitrate OPUS settings. The degradation at lower bitrates can be explained by the distortion introduced in the speech characteristics by the codec and removal of useful speech information.
The WER with the uncompressed test data is our baseline; for the analysis we use the audio from the same test set and encode it with 8kbps, 16kbps, 32kbps, and 128kbps per channel. Table 1 shows the relative word error rate (WER) degradation with the different bit rate audio inputs for the multi-channel model. As our results show, the WER degrades sharply after 16kbps, this is expected as the performance with the 8kbps encoding degrades drastically since OPUS encodes the audio as narrowband at 8kbps. With a bitrate of 128kbps per channel, the degradation reduces to 2.4% relative, demonstrating that most of the speech characteristics required for ASR are preserved in this case. We also visualize the spectrogram of a single audio example with various bitrates in Figure 2. As it can be seen from the figure, 8kbps encoding has no information above 4khz. The information loss, specially at higher frequencies is clear in both the 16kbps and 128kbps encoding.
Bitrate per channel | relative WER degradation % |
---|---|
Uncompressed | - |
8 | 202.1 |
16 | 12.6 |
32 | 6.6 |
128 | 2.4 |
Total bitrate | Bitrate for channel 1 | Bitrate for channel 2 | Relative WER degradation | ||
---|---|---|---|---|---|
Overall | Clean | Noisy | |||
Uncompressed | Uncompressed | Uncompressed | - | - | - |
264 | 256 | 8 | 11.1 | 10.5 | 10.9 |
264 | 132 | 132 | 2.7 | 3.7 | 2.2 |
252 | 256 | 16 | 4.1 | 3.7 | 4.4 |
252 | 136 | 136 | 2.7 | 3.7 | 2.2 |
288 | 256 | 32 | 1.5 | 1.9 | 1.1 |
288 | 144 | 144 | 2.7 | 3.7 | 2.2 |
320 | 256 | 64 | 0 | 0 | 0 |
320 | 160 | 160 | 2.8 | 4.3 | 2.2 |
384 | 256 | 128 | 0 | 0 | 0 |
384 | 192 | 192 | 2.4 | 3.7 | 2.2 |
We compare this degradation with similar inputs to a single channel model in Figure 3. For a fair comparison, we compare the WER for the single channel with bitrate with the multi-channel model with bitrate per channel, since the number of available bits for encoding the data is equal in those cases. The single channel model does not have the beamforming and the elastic spatial filtering layers. The rest of the components are initialized and trained similar to the multi-channel model . For this analysis, the baseline is the single channel model trained with uncompressed audio as input. The multi-channel model outperforms the single channel model for all bitrates by 4-6% relative depending on the total bits available for encoding. Comparing the performance at 256kbps for the single channel model and 64kbps for the multi-channel model, we note that the multi-channel model outperforms the single channel model by 1.52% relative. From these results, we infer that the OPUS compression preserves sufficient phase information from the waveforms to improve performance using the learned beamformer. The take away from this study is that given limited bandwidth, we should always prefer dividing the bandwidth into multiple channels instead of using all of it for encoding a single channel. This study demonstrates that we can potentially leverage from the multi-channel model architecture when the inputs are from different devices with different , in which case performing beamforming on a single device is not practical.
4 Distributing limited bandwidth among multiple channels
In Section 3, we have established that we should distributed limited bandwidth to multiple channels instead of a single channel. For the second part of the study, we wanted to understand how to distribute limited bandwidth between multiple channels. For this experiment, we construct two different test sets. For the first set of experiments, we fix the input to one of the channels to be the uncompressed audio and vary the bitrate of the second channel to 8,16,32, 64 and 128kbps. The second set is constructed where the kilo-bytes per second available is uniformly distributed between the 2 channels. For the purpose of the experiment, we assume that the uncompressed audio is equivalent to audio with a 256kbps rate, since our experiments showed that audio compressed with 256kbps produced the same WER as uncompressed audio. The baseline for this experiment is uncompressed audio for both channels as input. Our results are shown in Table 2.
From the results, we can see that given limited bandwidth, it is preferable to have a single channel with higher bitrate than to divide the bandwidth equally, given the other channel has a bitrate higher than 16kbps. We further analyze these results by noisy and clean conditions. We define the noisy condition as the SNR 5dB and clean conditions as SNR 5dB given the far-field nature of the audio. These results show that in both noisy and clean conditions, having a single channel encoded with the highest bit rate is beneficial over distributing the bitrate between the two channels equally. This demonstrates that the learned beamformer performs better when there is at least one channel with no degradation rather than both channels having some degradation, even when the bitrate is as low as 32kpbs. In fact we can recover performance when the bitrate of one channel is greater than 64kbps as long as we retain high bitrate on the other channel. However when the bitrate for one channel gets as low as 8kbps, its preferable to distribute the bitrate between the two channels evenly.
5 Mixed bitrate training
Next, we wanted to explore if the degradation can be compensated by matching the training data with the test data. For this purpose, we did 2 different sets of experiments. The first was to train individual multi-channel models for each bit-rate encoding we analyzed to see if we can compensate for the degradation . The second experiment was focused towards understanding if a single acoustic model can work well for different bitrates for different channels. We trained a model with mixed bitrates; for each training example, we picked an encoding bitrate uniformly from the set . In order to keep the comparison fair, we used the same number of training examples for training the mixed bitrate model and did not sample any training example multiple times with different bitrates. For each of these models, the acoustic scale was tuned on the development set that matched the training set.
Table 3 shows the results of these experiments. As expected, the training with matched bitrate encoding shows improvement over the baseline. For 32 kbps, the word error rate degradation reduces by 50%. The degradation with 16kbps can be attributed to the loss of information in encoding that cannot be recovered by matching the test data. The single mixed bitrate model can recover some of the degradation for all the different bitrate encodings we studied. For bitrates higher than 16, it is not as good as the matched training but better than the baseline. It is interesting to note however that for 16kbps the model can perform better than the matched training. We posit that this is because the matched model doesn’t see critical information for higher frequencies during training and hence cannot recover the errors. We also note that with this single model, we see a 3.4% degradation in WER when the audio input is uncompressed. This can be explained by the mismatch in the training and test data. This degradation can be recovered if the each training sample presented to the model with multiple encoding bitrates.
Test Bitrate per channel | Training data | WERR % |
Uncompressed | Uncompressed | - |
16 | Uncompressed | 12.6 |
16 | 16 | 10.0 |
16 | mixed bitrates | 7.9 |
Uncompressed | Uncompressed | - |
32 | Uncompressed | 6.6 |
32 | 32 | 3.3 |
32 | mixed bitrates | 4.1 |
Uncompressed | Uncompressed | - |
128 | Uncompressed | 2.4 |
128 | 128 | 1.6 |
128 | mixed bitrates | 2.1 |
Uncompressed | Uncompressed | - |
Uncompressed | mixed bitrates | 3.4 |
6 Analysis and Conclusions
In this work, we studied the degradation of WER as a function of bitrates available for the input audio. We showed that even with low bitrate encoding, the multi-channel model can outperform the single channel model, which opens up multi-channel modeling with multiple devices that may not support the same bitrate. Prior work with multi-device acoustic modeling has not focused on the mixed bitrate input [18][19]. We also demonstrated that when limited bandwidth is available, we should always choose a multi-channel input over a single channel input. For the best performance, it is optimal to encode one channel with the highest bandwidth available given the second channel has sufficient bandwidth available. Finally we demonstrate that by training the model with variable bitrates, we can potentially use a single model for inputs with all bitrates which is important for practical applications.
For our future work, we would like to explore an attention based mechanism to replace the beamformer, similar to the work the authors have proposed in [18] . This can be especially useful with mixed bitrate inputs so that the model can leverage the best information required from both channels for the task. In the future, we would also like to extend our study to using real data from multiple devices.
We would like to acknowledge the support of our colleague, Kenichi Kumatani, for this work.
References
- [1] K. Kumatani, M. Wu, S. Sundaram, N. Ström, and B. Hoffmeister, “Multi-geometry spatial acoustic modeling for distant speech recognition,” in Int. Conf. Acoustics, Speech and Signal Processing (ICASSP). IEEE, 2019, pp. 6635–6639.
- [2] T. N. Sainath, R. J. Weiss, K. W. Wilson, B. Li, A. Narayanan, E. Variani, M. Bacchiani, I. Shafran, A. Senior, K. Chin, et al., “Multichannel signal processing with deep neural networks for automatic speech recognition,” Transactions Audio, Speech, and Language Processing (TASLP), vol. 25, no. 5, pp. 965–979, 2017.
- [3] Dan Chazan, Gilad Cohen, Ron Hoory, and Meir Zibulski, “Low bit rate speech compression for playback in speech recognition systems,” in 2000 10th European Signal Processing Conference. IEEE, 2000, pp. 1–4.
- [4] Péter Orosz, Tamás Skopkó, Zoltán Nagy, and Tamás Lukovics, “Performance analysis of the opus codec in voip environment using qoe evaluation,” in The Eighth International Conference on Systems and Networks Communications, ICSNC, 2013.
- [5] Ingo Siegert, Alicia Flores Lotz, Linh Linda Duong, and Andreas Wendemuth, “Measuring the impact of audio compression on the spectral quality of speech data,” 2016.
- [6] Khoi-Nguyen C Mac, Xiaodong Cui, Wei Zhang, and Michael Picheny, “Large-scale mixed-bandwidth deep neural network acoustic modeling for automatic speech recognition,” arXiv preprint arXiv:1907.04887, 2019.
- [7] N García, JC Vásquez-Correa, JD Arias-Londoño, JF Várgas-Bonilla, and JR Orozco-Arroyave, “Automatic emotion recognition in compressed speech using acoustic and non-linear features,” in 2015 20th Symposium on Signal Processing, Images and Computer Vision (STSIVA). IEEE, 2015, pp. 1–7.
- [8] A Albahri, M Lech, and E Cheng, “Effect of speech compression on the automatic recognition of emotions,” International Journal of Signal Processing Systems, vol. 4, no. 1, pp. 55–61, 2016.
- [9] Ingo Siegert, Alicia Flores Lotz, Michael Maruschke, Oliver Jokisch, and Andreas Wendemuth, “Emotion intelligibility within codec-compressed and reduced bandwidth speech,” in Speech Communication; 12. ITG Symposium. VDE, 2016, pp. 1–5.
- [10] M. Wu, K. Kumatani, S. Sundaram, N. Ström, and B. Hoffmeister, “Frequency domain multi-channel acoustic modeling for distant speech recognition,” in Int. Conf. Acoustics, Speech and Signal Processing (ICASSP). IEEE, 2019, pp. 6640–6644.
- [11] Jean-Marc Valin, Koen Vos, and Timothy Terriberry, “Definition of the opus audio codec,” IETF, September, 2012.
- [12] Kota Solomon Raju and Abhinav Sharma, “Comparison of two speech communication codecs for transmitting voice/speech over zigbee,” in 2015 2nd International Conference on Signal Processing and Integrated Networks (SPIN). IEEE, 2015, pp. 685–690.
- [13] Arun Narayanan, Ananya Misra, Khe Chai Sim, Golan Pundak, Anshuman Tripathi, Mohamed Elfeky, Parisa Haghani, Trevor Strohman, and Michiel Bacchiani, “Toward domain-invariant speech recognition via large scale training,” in 2018 IEEE Spoken Language Technology Workshop (SLT). IEEE, 2018, pp. 441–447.
- [14] Ivan Himawan, Iain McCowan, and Sridha Sridharan, “Clustered blind beamforming from ad-hoc microphone arrays,” Transactions Audio, Speech, and Language Processing (TASLP), vol. 19, no. 4, pp. 661–676, 2010.
- [15] S. Doclo and M. Moonen, “Superdirective beamforming robust against microphone mismatch,” Transactions Audio, Speech, and Language Processing (TASLP), vol. 15, no. 2, pp. 617–631, 2007.
- [16] Karel Veselỳ, Arnab Ghoshal, Lukás Burget, and Daniel Povey, “Sequence-discriminative training of deep neural networks.,” in Interspeech, 2013, vol. 2013, pp. 2345–2349.
- [17] Koen Vos, Karsten Vandborg Sørensen, Søren Skak Jensen, and Jean-Marc Valin, “Voice coding with opus,” in Audio Engineering Society Convention 135. Audio Engineering Society, 2013.
- [18] Xiaofei Wang, Ruizhi Li, Sri Harish Mallidi, Takaaki Hori, Shinji Watanabe, and Hynek Hermansky, “Stream attention-based multi-array end-to-end speech recognition,” in ICASSP 2019-2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, 2019, pp. 7105–7109.
- [19] Samik Sadhu, Ruizhi Li, and Hynek Hermansky, “M-vectors: Sub-band based energy modulation features for multi-stream automatic speech recognition,” in ICASSP 2019-2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, 2019, pp. 6545–6549.