This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Three-class Overlapped Speech Detection using a
Convolutional Recurrent Neural Network

Abstract

In this work, we propose an overlapped speech detection system trained as a three-class classifier. Unlike conventional systems that perform binary classification as to whether or not a frame contains overlapped speech, the proposed approach classifies into three classes: non-speech, single speaker speech, and overlapped speech. By training a network with the more detailed label definition, the model can learn a better notion on deciding the number of speakers included in a given frame. A convolutional recurrent neural network architecture is explored to benefit from both convolutional layer’s capability to model local patterns and recurrent layer’s ability to model sequential information. The proposed overlapped speech detection model establishes a state-of-the-art performance with a precision of 0.6648 and a recall of 0.3222 on the DIHARD II evaluation set, showing a 20% increase in recall along with higher precision. In addition, we also introduce a simple approach to utilize the proposed overlapped speech detection model for speaker diarization which ranked third place in the Track 1 of the DIHARD III challenge.

Index Terms: overlapped speech detection, convolutional recurrent neural network, speaker diarization, deep learning

1 Introduction

Overlapped speech detection (OSD) is a task that estimates onsets and offsets of segments (i.e., a small part of an audio clip) within an audio clip (i.e., utterance, session, conversation as a whole) where more than one speaker is speaking simultaneously. OSD can be used for various applications, among which is speaker diarization where OSD is reported to play a critical role [1, 2, 3, 4, 5, 6, 7, 8, 9]. Without an OSD system, the performance of a speaker diarization system degrades, proportionally to the existing overlaps in a dataset, because the model conclusively misses the second speaker in the overlapped speech region.

Before the recent advances in deep neural networks (DNNs), hidden Markov models and Gaussian mixture models were widely used to build OSD systems [10, 11]. Martin et al. shows that the hidden Markov model can successfully perform OSD where each state is defined using a Gaussian mixture model. In recent works, OSD is treated as a sequence labeling task where convolutional neural networks (CNNs) or recurrent neural networks (RNNs) are trained as a binary classifier [12, 5, 13, 14]. Bullock et al. [14] adopts two bidirectional long short-term memory recurrent layers [15] and demonstrates state-of-the-art performance on the DIHARD II evaluation set.

In this study, we make two proposals on top of existing OSD researches. First, we train the OSD model as a three-class classifier instead of a binary classifier. Before the widespread use of DNNs in OSD, hidden Markov models were used for OSD[9] where OSD was modeled via transitions between three independent states: non-speech, single speaker speech, and overlapped speech. However, when using DNNs for OSD, only two classes (overlapped and non-overlapped) are defined. We argue that the DNN can also benefit from defining more specific classes based on additional motivations addressed in Section 2.1. Second, we use a convolutional recurrent neural network (CRNN) architecture to merit from both architecture’s features. Although studies adopting CRNN show state-of-the-art performance in other sequence labeling tasks such as sound event detection [16], existing OSD models adopt variants of either a CNN or a RNN. We assume that OSD models can be also improved by adopting a CRNN architecture, benefiting from both CNN and RNN’s unique modeling capability. In the upcoming sections, we explain and experimentally show that both proposals increase the performance of an OSD model and establish state-of-the-art performance.

Exploiting the developed OSD model, we also introduce a simple algorithm for detecting simultaneously speaking speakers in a diarization task. It adds two sub-processes using an OSD model to the speaker diarization pipeline. First, we add a sub-process that further segments speech activity detection (SAD) module’s output where the onsets and offsets of overlapped segments are adjusted using the SAD result. Second, we provide an additional label to the overlapped segment using the second ranked speaker based on the clustering result of speaker diarization. This method works independently from the mainstream speaker diarization pipeline, making it simple to utilize yet effective, ranking third in the DIHARD III challenge Track 1 core evaluation [17].

Refer to caption
Figure 1: Overall pipeline of the proposed CRNN-based three-class OSD system in the evaluation phase. An input audio clip is frame-wisely classified into one of the three classes where 0,1 and 2 represent non-speech, single speaker speech, and overlapped speech, respectively. To process the entire audio clip longer than the defined sequence length of 1.5 seconds, every 150 frames are input to the model with 50 frames shift and model outputs are combined.

2 Overlapped speech detection

Defined as a sequence labeling task, an OSD model inputs 𝐱=[𝐱𝟏,𝐱𝟐,,𝐱𝐭],𝐱𝐭f\mathbf{x}=[\mathbf{x_{1}},\mathbf{x_{2}},...,\mathbf{x_{t}}],\mathbf{x_{t}}\in\mathbb{R}^{f} where tt is the length of input sequence and ff refers to the dimensionality of each element. Acoustic features (e.g., mel-spectrograms or mel frequency cepstral coefficients) are widely chosen as the input feature. Corresponding label 𝐲=[y1,y2,,yt],yt\mathbf{y}=[y_{1},y_{2},...,y_{t}],y_{t}\in\mathbb{R} is used to guide the training where 0 refers to not overlapped and 1 refers to overlapped (can be the other way). The output layer of the model is either one node, trained using a binary cross-entropy loss function, or two nodes, trained using a categorical cross-entropy loss function.

The overall framework of our proposed OSD model when conducting inference using an input audio clip is illustrated in Figure 1. Note that for conventional DNN-based OSD models, labels are either 0 or 1 as mentioned earlier, but for the proposed three-class OSD model, labels are either 0, 1, or 2 indicating non-speech, single speaker speech, and overlapped speech. We adopt mel-spectrograms as the input feature, using a 25ms window and 10ms shift size. The OSD model inputs 150 frames (\approx 1.5s) at a time with 50 frames shift. The node of the output layer that indicates overlapped speech is used. Except for the first and the last 100 frames, the sum of model output is divided by three accounting for the multiple predictions caused by shifting the input. To compensate for the decrease in the length of sequence affected by pooling layers, the values are duplicated. The main contributions of this paper, three-class classification and the CRNN architecture, are described in Sections 2.1 and 2.2 respectively.

Table 1: Architecture of the proposed CRNN for three-class OSD. A mel-spectrogram of size (150, 128, 1) is input to the DNN where 150 refers to the length of sequence (1.5s), 128 is the number of mel bins, and 1 is the number of channels. Two numbers in the Conv bracket refers to filter length and stride size (BN: batch normalization, SE: squeeze-excitation). Average is applied to the mel bin dimension, before the GRU layer.
Layer Input:(#seq,#mel,#filt) Output shape
CNN {Conv(3,1)BNReLUConv(3,1)BNReLUSEAveragePool}\left\{\begin{tabular}[]{c}Conv(3,1)\\ BN\\ ReLU\\ Conv(3,1)\\ BN\\ ReLU\\ SE\\ AveragePool\\ \end{tabular}\right\} ×\times3 (25, 32, 128)
Average Average (25,128)
GRU bi-GRU(256)×\times2 (512,)
Linear {FC(256)Dropout(0.5)LeakyReLULinear(3)}\left\{\begin{tabular}[]{c}FC(256)\\ Dropout(0.5)\\ LeakyReLU\\ Linear(3)\\ \end{tabular}\right\} (3,)

2.1 OSD as a three-class classification

Conventional OSD systems using DNNs are trained to perform binary classification because the objective at the inference phase is to decide whether overlapped speech exists for each frame. It is a widely accepted straightforward approach, training the model to perform as one wants. However, by training the OSD model as a two-class classifier, both non-speech and single speaker segments are labeled as not overlapped. We assume that grouping non-speech and single speaker speech into not overlapped, regardless of the existence of human voice, may not be ideal in two aspects. First, two small clusters may be created within not overlapped class, diminishing the OSD model’s generalization. Second, training OSD as a two-class classification task intensifies class imbalance. For instance, suppose there exists a sample in the train set that has 40 non-speech frames, 100 single speaker speech frames, and 10 overlapped speech frames. Training it with conventional two-class results in 14:1 imbalance ratio, but with three-class training, it becomes 4:10:1. Although manual weight can be set for each class, it is difficult to account for the single speaker segments.

Thus, we propose to train the OSD model as a three-class classifier where the classes are non-speech, single speaker speech, and overlapped speech. This is inspired by studies [10, 11] that use hidden Markov models where single speaker speech and non-speech are separately defined as states. By dividing not overlapped class into single speaker speech and non-speech classes, we argue that not only the problem occurred by grouping different categories may decrease, but also the DNN can leverage from additional information. It has commonalities in terms of motivation with the multitask learning [18] framework where more specific and concrete information is additionally provided to help learning. Also, precise class balancing using loss weights becomes possible. The proposed three-class scheme has no additional requirements when performing inference. One can simply use the node of the output layer that indicates overlapped speech (before or after applying softmax non-linearity) as the model’s score for the given input frame.

2.2 CRNN architecture

Most preceding studies that use DNN-based models for the OSD task either utilize a CNN or a RNN [12, 5, 13, 14]. However, we propose to use a CRNN architecture with three convolution blocks and two gated recurrent unit (GRU) layers. Three motivations led us to propose and use the CRNN architecture. First, CNNs and RNNs have different aspects that the OSD model can benefit from. CNNs are known for their capability of modeling local patterns and RNNs are widely used for their capability of modeling sequential data. We argue that by using a CRNN architecture, advantages of both CNNs and RNNs can be exploited. Second, decreasing the length of sequence via pooling layers included in CNN will facilitate the training of GRU layers. In addition, decreasing the sequence length will give a smoothing effect on the model’s output. For example, according to our configuration when 150 frames are input, the CRNN model will output 25 frames where each output frame represents six input frames. Third, a CRNN architecture shows state-of-the-art performance in sound event detection [16] which is also a sequence labeling task that predicts onsets and offsets of various predefined sound events.

Table 1 denotes the proposed CRNN architecture used for OSD. Three convolution blocks are used where the length of a sequence is divided by 6, followed by an average on the mel bin dimension. A squeeze-excitation [19] is applied to each block. Details on how the length of the sequence and the number of mel bins are decreased are addressed in Section 4.2. The output of CNN is input to two bidirectional GRU layers that output sequences. Each frame is then classified among the three classes and lastly duplicated to match the length of the input sequence.

Refer to caption
Figure 2: Further segmentation of SAD segment using OSD result.

2.3 Overlap and sampling rate augmentation

We adopt two data augmentation techniques on top of widely used noise and reverberation augmentation [20, 21]. First, we generate augmented overlap speech samples using two single speaker segments alike the method introduced in Bullock et al. [14]. It diminishes the class imbalance problem and generates sufficient overlapped speech samples; most datasets that can be used for training an OSD model have less than 10% overlapped speech regions. Second, we augment training samples by downsampling an audio clip and then upsampling back. We use this augmentation to cope with telephone conversation audio clips that have been upsampled in the test phase. Through experiments, we empirically find that this augmentation mitigates performance degradation on 8kHz sampling rate audio clips that exist in the evaluation dataset.

3 Overlapped segment clustering in speaker diarization

Speaker diarization is a widely used application of an OSD model. Processing overlapped segments have been recognized as a key for developing state-of-the-art speaker diarization systems. This is because, in the conventional diarization systems that cannot assign multiple labels to one segment, the ratio of overlapped speech directly leads to an increase in missed speech, increasing diarization error rate (DER).

In order to utilize the proposed OSD model for speaker diarization, we introduce two simple steps that can be applied on top of the existing speaker diarization pipeline. First, we use the OSD model to further divide SAD results into single speaker segment and overlapped speech segment. Regions where both the OSD model and the SAD model detects are assigned as overlapped speech segment. Figure 2 illustrates three possible cases of the OSD model further segmenting the SAD segment. If the OSD onset is earlier than the SAD onset, SAD onset is used as the onset of an overlapped segment and OSD offset is used as modified SAD segment onset (Figure 2-(a)). If the OSD offset positions later than the SAD offset, SAD offset is used as an overlapped segment’s offset and SAD offset is replaced by OSD onset (Figure 2-(b)). When the OSD segment exists in the middle of an SAD segment, it is divided into three segments (Figure 2-(c)).

The other application of the OSD model for speaker diarization occurs after the clustering phase, in a similar manner with [14]. When performing clustering, there exists a speaker posterior matrix where speakers are aligned from the most likely to least likely for each segment. For overlapped segments, we simply allocate the second most likely speaker as an additional label. This approach is simple, easy to implement, and can be applied to the result of various clustering methods (e.g., spectral and agglomerative hierarchical clustering [22, 23, 24]) that can derive the second most likely speaker for each segment. However, note that the limitation of this approach is not being able to detect when three or more speakers are uttering simultaneously.

Aforementioned approach is applied to the Clova team’s speaker diarization model [25]. This model uses a reference SAD following the rules of Track 1 of the DIHARD III challenge. For speaker embedding extraction, a variant of ResNet34V2 in voxceleb_trainer111https://github.com/clovaai/voxceleb_trainer is used. Speaker embeddings are then fed into two feature enhancement modules, followed by spectral clustering and agglomerative hierarchical clustering [22, 23]. Further information regarding this speaker diarization system including feature enhancement techniques can be found in [21].

4 Experiments

4.1 Datasets

We use four datasets to train the proposed OSD model: AMI Meeting corpus [26], DIHARD I development set[27], DIHARD II development set [28], and VoxConverse dataset [29]. The AMI corpus includes 100 hours of meeting conversations. DIHARD I and II development sets include 19 and 23 hours of data from diverse domains respectively. The VoxConverse dataset consists of approximately 20 hours of conversation from YouTube videos. For all datasets, multi-channel audio clips were modified to single-channel and the sample rate of all audio clips was adjusted to 16kHz. The DIHARD II evaluation set and the DIHARD III development set [30] are used to report the performance of the proposed OSD model. OSD model’s performance is reported using precision and recall. Speaker diarization result is reported in DER, the summation of false alarm, miss, and confusion.

4.2 Configurations and details

We use a 128-dimensional log mel-spectrograms (i.e., filterbank energies) as input to the DNN with a window size of 25ms and 10ms shift. Pre-emphasis is applied before applying Fourier transforms and mean normalization is applied to the extracted log mel-spectrograms.

Data augmentation. A copy of all audio clips with a sampling rate of 16kHz is augmented by downsampling to 8kHz and upsampling back. We also apply data augmentation using reverberation and noise from RIRs and Musan datasets [31, 32], following existing recipes [20, 21]. Overlapped speech sample augmentation following the recipe introduced in Bullock et al. [14] is also applied.

DNN training and inference. We modify the input sample to 150 frames (1.5s) for batch construction when training the CRNN model. An average pooling with size (2,1), (3,2), (1,2) is applied after each convolution block in Table 1, decreasing the length of input sequence and number of frequency bins to 25 frames and 32, respectively. Each output frame is six times duplicated to match the original sequence length. The model is trained 60 epochs using a batch size of 1024, using four Nvidia Tesla P40 GPUs. Adam optimizer with cosine annealing learning rate scheduler is used. To account for class imbalance, we give weights to categorical cross-entropy loss proportionally. At the inference phase, we input 150 frames to the model with 50 frames shift to handle audio clips with varying duration.

Table 2: Comparison with the reported state-of-the-art OSD model on DIHARD II evaluation set. Performance of the identical model is additionally reported on DIHARD III development set as a reference (DH: DIHARD, higher is better for both precision and recall).
System Dataset Precision Recall
Bullock et al.[14] DH2 eval 0.6450 0.2670
Ours DH2 eval 0.6648 0.3222
Ours DH3 dev 0.9000 0.4609
Table 3: Comparison experiments with ablations showing the effectiveness of the proposed three-class OSD approach and CRNN architecture. Threshold was set to a point where the system shows precision of 0.9 (higher is better for both precision and recall).
System Precision Recall
Best (all combined) 0.9000 0.4609
w/o three-class training 0.9000 0.4113
w/o CNN 0.9000 0.4318
w/o GRU 0.9000 0.3603
3 GRU layers 0.9000 0.4365
1 GRU layers 0.9000 0.3314
w/o sequence compression 0.9000 0.4471
13\frac{1}{3} sequence compression 0.9000 0.4397
110\frac{1}{10} sequence compression 0.9000 0.4135

4.3 Results

Table 2 describes the comparison between the existing model and the proposed OSD model applying both three-class training and CRNN architecture. The OSD model proposed in Bullock et al. [14], which demonstrated state-of-the-art performance in AMI and DIHARD II evaluation set is used as the baseline. Experimental results show that our proposed OSD model establishes a new state-of-the-art performance on the DIHARD II evaluation set. As the DIHARD III challenge has been held recently, we further report the proposed model’s performance on its development set. Note that this model is similar to, but performs slightly better than the OSD model we used when participating the DIHARD III challenge [25]. Hereafter, we report all performances using the DIHARD III development set.

Table 3 addresses comparison experiments with ablation studies to show the effectiveness of training the OSD model as a three-class classifier and using a CRNN architecture. Using the best performing system in the first row as a pivot, we first remove the individual proposed approach. Then, ablations experiments on the number of GRU layers and the extent of sequence length compression are dealt. All performances are reported using a threshold that shows a precision of 0.9.

Results from the second to the fourth rows show that both three-class training and CRNN architecture successfully improves the OSD model’s performance. Thee-class training relatively improved recall by 12%. CRNN architecture was also proven effective where removing GRU layers degraded the performance more. Through the fifth and sixth rows, we conclude that the model benefits from multiple GRU layers where using one GRU layer decreased performance the most. The last three rows show that compressing the sequence length using pooling layers, before GRU, is also effective for OSD. Reducing training sequence length was effective than not reducing at all, but significantly decreasing also decreased the performance.

Table 4 reports speaker diarization performance using the proposed OSD model on the DIHARD III Track 1 development set. Performance is reported using DER, equal to the summation of false alarm, miss, and confusion. The speaker diarization system refers to that of the Naver Clova team’s submission for the DIHARD III challenge [25]. OSD model threshold was set to a value that shows a precision of 0.95.

Experimental results show that further segmentation and allocating additional speaker labels using the developed OSD model successfully lowers the DER of speaker diarization, decreasing the DER from 14.29% to 13.74% on the DIHARD III development set. Miss, which the OSD model should reduce, decreased to 9.40% showing relative improvement of 8.73%, at the cost of an increase in false alarm by 0.09 which was caused by the OSD model predicting single speaker regions as overlapped. Our system shows 32% lower DER than the challenge baseline and are comparable to the challenge winning system.

Table 4: Effectiveness of the OSD model when applied to the speaker diarization pipeline using the DIHARD III development set track 1 configuration (FA: false alarm, Conf: confusion, lower is better for all four metrics).
OSD configuration DER FA Miss Conf
w/o OSD 14.29 0.43 10.30 3.56
w/ OSD 13.74 0.52 9.40 3.81
Challenge Baseline [33] 20.25 - - -
Challenge Winner [34] 13.30 - - -

5 Conclusion

We proposed an OSD model that is trained as a three-class classifier dividing the non-overlapped class into non-speech and single speaker speech classes. A CRNN architecture was also proposed to incorporate both CNN and RNN’s capability in detecting overlapped speech. Combining two proposals, our OSD model demonstrates state-of-the-art performance on the DIHARD II evaluation set. An approach of using a developed OSD model for speaker diarization was also addressed, where the system that followed this approach ranked third place in DIHARD III Track 1 challenge.

References

  • [1] K. Boakye, B. Trueba-Hornero, O. Vinyals, and G. Friedland, “Overlapped speech detection for improved speaker diarization in multiparty Meetings,” in International Conference on Acoustics, Speech and Signal Processing (ICASSP), 2008, pp. 4353–4356.
  • [2] D. Charlet, C. Barras, and J. S. Lienard, “Impact of overlapping speech detection on speaker diarization for broadcast news and debates,” in International Conference on Acoustics, Speech and Signal Processing (ICASSP), 2013, pp. 7707–7711.
  • [3] O. Ben-Harush, H. Guterman, and I. Lapidot, “Frame level entropy based overlapped speech detection as a pre-processing stage for speaker diarization,” in Machine Learning for Signal Processing (MLSP), 2009.
  • [4] K. Boakye, O. Vinyals, and G. Friedland, “Two’s a Crowd: Improving Speaker Diarization by Automatically Identifying and Excluding Overlapped Speech,” in Annual Conference of the ISCA, INTERSPEECH, 2008.
  • [5] M. Yousefi and J. H. Hansen, “Frame-Based Overlapping Speech Detection Using Convolutional Neural Networks,” in International Conference on Acoustics, Speech and Signal Processing (ICASSP), 2020, pp. 6744–6748.
  • [6] V. Andrei, H. Cucu, and C. Burileanu, “Detecting overlapped speech on short timeframes using deep learning,” in Annual Conference of the International Speech Communication Association, INTERSPEECH, 2017, pp. 1198–1202.
  • [7] M. Diez, F. Landini, L. Burget, J. Rohdin, A. Silnova, K. Zmolíková, O. Novotný, K. Veselý, O. Glembek, O. Plchot, L. Mošner, and P. Matějka, “BUT system for DIHARD speech diarization challenge 2018,” in Annual Conference of the International Speech Communication Association, INTERSPEECH, 2018, pp. 2798–2802.
  • [8] S. H. Yella and H. Bourlard, “Overlapping speech detection using long-term conversational features for speaker diarization in meeting room conversations,” IEEE/ACM Transactions on Audio Speech and Language Processing, vol. 22, no. 12, pp. 1688–1700, 12 2014.
  • [9] S. N. Wrigley, G. J. Brown, V. Wan, and S. Renals, “Speech and crosstalk detection in multichannel audio,” IEEE Transactions on Speech and Audio Processing, vol. 13, no. 1, pp. 84–91, 1 2005.
  • [10] Z. Martin and H. Javier, “The Detection of Overlapping Speech with Prosodic Features for Speaker Diarization,” in Annual Conference of the ISCA, INTERSPEECH, 2011.
  • [11] O. Ben-Harush, L. Itshak, and G. Hugo, “Entropy based overlapped speech detection as a pre-processing stage for speaker diarization.” in Annual Conference of the ISCA, INTERSPEECH, 2009.
  • [12] S. Cornell, M. Omologo, S. Squartini, and E. Vincent, “Detecting and counting overlapping speakers in distant speech scenarios,” in Annual Conference of the ISCA, INTERSPEECH, 2020, pp. 3107–3111.
  • [13] J. T. Geiger, F. Eyben, B. Schuller, and G. Rigoll, “Detecting Overlapping Speech with Long Short-Term Memory Recurrent Neural Networks,” in Annual Conference of the ISCA, INTERSPEECH, 2013.
  • [14] L. Bullock, H. Bredin, and L. P. Garcia-Perera, “Overlap-Aware Diarization: Resegmentation Using Neural End-to-End Overlapped Speech Detection,” in International Conference on Acoustics, Speech and Signal Processing (ICASSP).   IEEE, 2020, pp. 7114–7118.
  • [15] S. Hochreiter and J. Schmidhuber, “Long Short-Term Memory,” Neural Computation, vol. 9, no. 8, pp. 1735–1780, 1997.
  • [16] L. Pi and X. Zheng, “Polyphonic Sound Event Detection and Localization Using a Two-Stage Strategy,” in DCASE, 2019.
  • [17] N. Ryant, K. Church, C. Cieri, J. Du, S. Ganapathy, and M. Liberman, “Third DIHARD Challenge Evaluation Plan,” arXiv, 6 2020.
  • [18] R. Caruana, “Multitask Learning,” Machine Learning, vol. 28, pp. 41–75, 1997.
  • [19] J. Hu, L. Shen, and G. Sun, “Squeeze-and-Excitation Networks,” in IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2018, pp. 7132–7141.
  • [20] D. Povey, G. Boulianne, L. Burget, P. Motlicek, and P. Schwarz, “The Kaldi Speech Recognition Toolkit,” in IEEE Automatic Speech Recognition and Understanding Workshop (ASRU), 2011.
  • [21] H. S. Heo, B. J. Lee, J. Huh, and J. S. Chung, “Clova baseline system for the VoxCeleb speaker recognition challenge 2020,” arXiv, 9 2020.
  • [22] V. U. Luxburg, “A tutorial on spectral clustering,” Statistics and Computing, vol. 17, no. 4, pp. 395–416, 12 2007.
  • [23] H. Ning, M. Liu, H. Tang, and T. S. Huang, “A Spectral Clustering Approach to Speaker Diarization,” in Annual Conference of the ISCA, INTERSPEECH, 2006.
  • [24] K. J. Han, S. Kim, and S. S. Narayanan, “Strategies to improve the robustness of agglomerative hierarchical clustering under data source variation for speaker diarization,” IEEE Transactions on Audio, Speech and Language Processing, vol. 16, no. 8, pp. 1590–1601, 11 2008.
  • [25] H.-S. Heo, J.-w. Jung, Y. Kwon, J. Kim, J. Huh, J. S. Chung, and B.-J. Lee, “NAVER CLOVA SUBMISSION TO THE THIRD DIHARD CHALLENGE,” Tech. Rep., 2021. [Online]. Available: https://dihardchallenge.github.io/dihard3/system_descriptions/dihard3_system_description_team73.pdf
  • [26] J. Carletta, “Unleashing the killer corpus: Experiences in creating the multi-everything AMI Meeting Corpus,” in Language Resources and Evaluation, vol. 41, no. 2.   Springer, 2007, pp. 181–190.
  • [27] N. Ryant, K. Church, C. Cieri, J. Du, S. Ganapathy, M. Liberman, and L. D. Consortium, “First DIHARD challenge evaluation plan,” Tech. Rep., 2018.
  • [28] N. Ryant, K. Church, C. Cieri, A. Cristia, J. Du, S. Ganapathy, and M. Liberman, “The second dihard diarization challenge: Dataset, task, and baselines,” in Annual Conference of the ISCA, INTERSPEECH, 2019, pp. 978–982.
  • [29] J. S. Chung, J. Huh, A. Nagrani, T. Afouras, and A. Zisserman, “Spot the conversation: Speaker diarisation in the wild,” in Annual Conference of the International Speech Communication Association, INTERSPEECH, 2020, pp. 299–303.
  • [30] N. Ryant, P. Singh, V. Krishnamohan, R. Varma, K. Church, C. Cieri, J. Du, S. Ganapathy, and M. Liberman, “The Third DIHARD Diarization Challenge,” arXiv, 12 2020.
  • [31] T. Ko, V. Peddinti, D. Povey, M. L. Seltzer, and S. Khudanpur, “A study on data augmentation of reverberant speech for robust speech recognition,” in International Conference on Acoustics, Speech and Signal Processing (ICASSP).   IEEE, 2017, pp. 5220–5224.
  • [32] D. Snyder, G. Chen, and D. Povey, “MUSAN: A Music, Speech, and Noise Corpus,” arXiv, 10 2015.
  • [33] N. Ryant, P. Singh, V. Krishnamohan, R. Varma, K. Church, C. Cieri, J. Du, S. Ganapathy, and M. Liberman, “The Third DIHARD Diarization Challenge,” Tech. Rep., 2021. [Online]. Available: https://dihardchallenge.github.io/dihard3/system_descriptions/dihard3_system_description_team18.pdf
  • [34] Y. Wang, M. He, S. Niu, L. Sun, T. Gao, X. Fang, J. Pan, J. Du, and C.-H. Lee, “USTC-NELSLIP System Description for DIHARD-III Challenge,” Tech. Rep., 2021. [Online]. Available: https://dihardchallenge.github.io/dihard3/system_descriptions/dihard3_system_description_team54.pdf