ICASSP 2021 Deep Noise Suppression Challenge
Abstract
The Deep Noise Suppression (DNS) challenge is designed to foster innovation in the area of noise suppression to achieve superior perceptual speech quality. We recently organized a DNS challenge special session at INTERSPEECH 2020. We open-sourced training and test datasets for researchers to train their noise suppression models. We also open-sourced a subjective evaluation framework and used the tool to evaluate and pick the final winners. Many researchers from academia and industry made significant contributions to push the field forward. We also learned that as a research community, we still have a long way to go in achieving excellent speech quality in challenging noisy real-time conditions. In this challenge, we are expanding both our training and test datasets. Clean speech in the training set has increased by 200% with the addition of singing voice, emotion data, and non-English languages. The test set has increased by 100% with the addition of singing, emotional, non-English (tonal and non-tonal) languages, and, personalized DNS test clips. There are two tracks with a focus on (i) real-time denoising, and (ii) real-time personalized DNS.
Index Terms— Speech Enhancement, Perceptual Speech Quality, P.808, Deep Noise Suppressor, Machine Learning.
1 Introduction
In recent times, remote work has become the ”new normal” as the number of people working remotely has exponentially increased due to the pandemic. There has been a surge in the demand for reliable collaboration and real-time communication tools. Audio calls with very good to excellent speech quality are needed during these times as we try to stay connected and collaborate with people every day. We are easily exposed to a variety of background noises such as a dog barking, a baby crying, kitchen noises, etc. Background noise significantly degrades the quality and intelligibility of the perceived speech leading to fatigue. Background noise poses a challenge in other applications such as hearing aids and smart devices.
Real-time Speech Enhancement (SE) for perceptual quality is a decades old classical problem and researchers have proposed numerous solutions [1, 2]. In recent years, learning-based approaches have shown promising results [3, 4, 5]. The Deep Noise Suppression (DNS) Challenge organized at INTERSPEECH 2020 showed promising results, while also indicating that we are still about 1.4 Differential Mean Opinion Score (DMOS) from the ideal Mean Opinion Score (MOS) of 5 when tested on the DNS Challenge test set [6, 7]. The DNS Challenge is the first contest that we are aware of that used subjective evaluation to benchmark SE methods using a realistic noisy test set [8]. We open sourced clean speech and noise corpus with configurable scripts to generate noisy-clean speech pairs suitable to train a supervised noise suppression model. There were two tracks, real-time and non-real-time based on the computational complexity of the inference. We received an overwhelming response to the challenge with participation from a diverse group of researchers, developers, students, and hobbyists from both academia and industry. We also received positive responses from the participants as many found the open sourced datasets quite useful, and both the dataset and test framework have been cloned at a fast rate since the challenge.
The ICASSP 2021 Deep Noise Suppression (DNS) Challenge111https://github.com/microsoft/DNS-Challenge is intended to stimulate research in the area of real-time noise suppression. For ease of reference, we will call the ICASSP 2021 challenge as DNS Challenge 2 and the Interspeech 2020 challenge as DNS Challenge 1. The DNS Challenge 2 will have a real-time denoising track similar to the one in DNS Challenge 1. In addition, we will have a personalized DNS track focused on using speaker information to achieve better perceptual quality. In addition to the datasets we open sourced for DNS Challenge 1, we increased clean speech in training set by 50% resulting in over 760 hours which includes singing voice, emotion data, and non-english languages (Chinese). Noise data in training set remains the same as DNS Challenge 1. We provide over 118,000 room impulse responses (RIR), which includes real and synthetic RIRs from public datasets. We provide acoustic parameters: Reverberation time (T60) and Clarity (C50) for read clean speech and RIR sample. In DNS Challenge 2, we increased the testset by 100% by adding emotion data, singing voice, non-English languages in Track 1 and real and synthetic clips for personalized DNS in Track 2. For DNS Challenge 1, we open sourced a subjective evaluation framework based on ITU-T P.808 [9]. The final evaluation of the participating models were done based on subjective evaluation using the P.808 subjective testing framework. We describe the results of the challenge at the end.
2 Challenge Tracks
The challenge had the following two tracks:
-
1.
Track 1: Real-Time Denoising track requirements
-
•
The noise suppressor must take less than the stride time (in ms) to process a frame of size (in ms) on an Intel Core i5 quad-core machine clocked at 2.4 GHz or equivalent processors. For example, for 50% overlap between frames. The total algorithmic latency allowed including the frame size , stride time , and any look ahead must be 40ms. For example, for a real-time system that receives 20ms audio chunks, if you use a frame length of 20ms with a stride of 10ms resulting in an algorithmic latency of 30ms, then you satisfy the latency requirements. If you use a frame size of 32ms with a stride of 16ms resulting in an algorithmic latency of 48ms, then your method does not satisfy the latency requirements as the total algorithmic latency exceeds 40ms. If your frame size plus stride is less than 40ms, then you can use up to ms future information.
-
•
-
2.
Track 2: Personalized Deep Noise Suppression (pDNS) track requirements
-
•
Satisfy Track 1 requirements.
-
•
You will have access to 2 minutes speech of a particular speaker to extract and adapt speaker related information that might be useful to improve the quality of the noise suppressor. The enhancement must be done on the noisy speech test segment of the same speaker.
-
•
The enhanced speech using speaker information must be of better quality than enhanced speech without using the speaker information.
-
•
3 Training Datasets
The goal of releasing the clean speech and noise datasets is to provide researchers with an extensive and representative dataset to train their SE models. We initially released MSSNSD [10] with a focus on extensibility, but the dataset lacked the diversity in speakers and noise types. We published a significantly larger and more diverse data set with configurable scripts for DNS Challenge 1 [8]. Many researchers found this dataset useful to train their noise suppression models and achieved good results. However, the training and the test datasets did not include clean speech with emotions such as crying, yelling, laughter or singing. Also, the dataset only includes the English language. For DNS Challenge 2, we are adding speech clips with other emotions and included about 10 non-English languages. Clean speech in training set is total 760.53 hours: read speech (562.72 hours), singing voice (8.80 hours), emotion data (3.6hours), Chinese mandarin data (185.41 hours). We have grown clean speech to 760.53 hours as compared to 562.72 hours in DNS Challenge 1. The details about the clean and noisy dataset are described in the following sections.
3.1 Clean Speech
Clean speech consists of three subsets: (i) Read speech recorded in clean conditions; (ii) Singing clean speech; (iii) Emotional clean speech; and (iv) Non-english clean speech. The first subset is derived from the public audiobooks dataset called Librivox222https://librivox.org/. It is available under the permissive creative commons 4.0 license [11]. It has recordings of volunteers reading over 10,000 public domain audiobooks in various languages, the majority of which are in English. In total, there are 11,350 speakers. Many of these recordings are of excellent speech quality, meaning that the speech was recorded using good quality microphones in a silent and less reverberant environments. But there are many recordings that are of poor speech quality as well with speech distortion, background noise, and reverberation. Hence, it is important to clean the data set based on speech quality. We used the online subjective test framework ITU-T P.808 [9] to sort the book chapters by subjective quality. The audio chapters in Librivox are of variable length ranging from few seconds to several minutes. We randomly sampled 10 audio segments from each book chapter, each of 10 seconds in duration. For each clip, we had 2 ratings, and the MOS across all clips was used as the book chapter MOS. Figure 1 shows the results, which show the quality spanned from very poor to excellent quality. In total, it is 562 hours of clean speech, which was part of DNS Challenge 1.
The second subset consists of high-quality audio recordings of singing voice recorded in noise-free conditions by professional singers. This subset is derived from VocalSet corpus [12] with Creative Commons Attribution 4.0 International License (CC BY 4.0). license. It has 10.1 hours of clean singing voice recorded by 20 professional singers: 9 males, and 11 females. This data was recorded on a range of vowels, a diverse set of voices on several standard and extended vocal techniques, and sung in contexts of scales, arpeggios, long tones, and excerpts. We downsampled the mono .WAV files from 44.1kHz to 16kHz and added it to clean speech used by the training data synthesizer.
The third subset consists of emotion speech recorded in noise-free conditions. This is derived from Crowd-sourced Emotional Mutimodal Actors Dataset (CREMA-D) [13] made available under the Open Database License. It consists of 7,442 audio clips from 91 actors: 48 male, and 43 female accounting to total 3.5 hours of audio. The age of the actors was in the range of 20 to 74 years with diverse ethnic backgrounds including African America, Asian, Caucasian, Hispanic, and Unspecified. Actors read from a pool of 12 sentences for generating this emotional speech dataset. It accounts for six emotions: Anger, Disgust, Fear, Happy, Neutral, and Sad at four intensity levels: Low, Medium, High, Unspecified. The recorded audio clips were annotated by multiple human raters in three modalities: audio, visual, and audio-visual. Categorical emotion labels and real-value emotion level values of perceived emotion were collected using crowd-sourcing from 2,443 raters. This data was provided as 16 kHz .WAV files so we added it to our clean speech as it is.
The fourth subset has clean speech from non-English languages. It consists of both tonal and non-tonal languages including Chinese (Mandarin), German and Spanish. Mandarin data consists of OpenSLR18 333http://www.openslr.org/18/ THCHS-30 [14] and OpenSLR33 444http://www.openslr.org/33/ AISHELL [15] datasets, both with Apache 2.0 license. THCHS30 was published by Center for Speech and Language Technology (CSLT) at Tsinghua University for speech recognition. It consists of 30+ hours of clean speech recorded at 16-bit 16 kHz in noise-free conditions. Native speakers of standard Mandarin read text prompts chosen from a list of 1000 sentences. We added the entire THCHS-30 data in our clean speech for the training set. It consisted of 40 speakers: 9 male, 31 female in the age range of 19-55 years. It has total 13,389 clean speech audio files [14]. The AISHELL dataset was created by Beijing Shell Shell Technology Co. Ltd. It has clean speech recorded by 400 native speakers ( 47% male and 53% female) of Mandarin with different accents. The audio was recorded in noise-free conditions using high fidelity microphones. It is provided as 16-bit 16kHz .wav files. It is one of the largest open-source Mandarin speech datasets. We added the entire AISHELL corpus with 141,600 utterances spanning 170+ hours of clean Mandarin speech to our training set.
Spanish data is 46 hours of clean speech derived from OpenSLR39, OpenSLR61, OpenSLR71, OpenSLR73, OpenSLR74 and OpenSLR75 where re-sampled all .WAV files to 16 kHz. German data is derived from four corpora namely (i) The Spoken Wikipedia Corpora [16], (ii) Telecooperation German Corpus for Kinect [17], (iii) M-AILABS data [18], (iv) zamia-speech forschergeist corpora. Complete German data constitute 636 hours. Italian (128 hours), French (190 hours), Russian (47 hours) are taken from M-AILABS data [18]. M-AILABS Speech Dataset is a publicly available multi-lingual corpora for training speech recognition and speech synthesis systems.

3.2 Noise
The noise clips were selected from Audioset 555https://research.google.com/audioset/ [19] and Freesound 666https://freesound.org/. Audioset is a collection of about 2 million human labeled 10s sound clips drawn from YouTube videos and belong to about 600 audio events. Like the Librivox data, certain audio event classes are over-represented. For example, there are over a million clips with audio classes music and speech and less than 200 clips for classes such as toothbrush, creak, etc. Approximately 42% of the clips have a single class, but the rest may have 2 to 15 labels. Hence, we developed a sampling approach to balance the dataset in such a way that each class has at least 500 clips. We also used a speech activity detector to remove the clips with any kind of speech activity, to strictly separate speech and noise data. The resulting dataset has about 150 audio classes and 60,000 clips. We also augmented an additional 10,000 noise clips downloaded from Freesound and DEMAND databases [20]. The chosen noise types are more relevant to VOIP applications. In total, there is 181 hours of noise data.
3.3 Room Impulse Responses
We provide 3076 real and approximately 115,000 synthetic rooms impulse responses (RIRs) where we can choose either one or both types of RIRs for convolving with clean speech. Noise is then added to reverberant clean speech while DNS models are expected to take noisy reverberant speech and produce clean reverberant speech. Challenge participants can do both de-reverb and denoising with their models if they prefer. These RIRs are chosen from openSLR26 [21] 777http://www.openslr.org/26/ and openSLR28 [21] 888http://www.openslr.org/28/ datasets, both released with Apache 2.0 License.
3.4 Acoustic parameters
We provide two acoustic parameters: (i) Reverberation time, T60 [22] and (ii) Clarity, C50 [23] for all audio clips in clean speech of training set. We provide T60, C50 and isReal Boolean flag for all RIRs where isReal is 1 for real RIRs and 0 for synthetic ones. The two parameters are correlated. A RIR with low C50 can be described as highly reverberant and vice versa [22, 23]. These parameters are supposed to provide flexibility to researchers for choosing a sub-set of provided data for controlled studies.
4 Test set
In DNS Challenge 1, the test set consisted of 300 real recordings and 300 synthesized noisy speech clips. The real clips were recorded internally at Microsoft and also using crowdsourcing tools. Some of the clips were taken from Audioset. The synthetic clips were divided into reverberant and less reverberant clips. These utterances were predominantly in English. All the clips are sampled at 16 kHz with an average clip length of 12 secs. The development phase test set is in the ”ICASSP_dev_test_set” directory in the DNS Challenge repository. For this challenge, the primary focus is to make the test set as realistic and diverse as possible.
4.1 Track 1
Similar to DNS Challenge 1, the test set for DNS Challenge 2 is divided into real recordings and synthetic categories. However, the synthetic clips are mainly composed of the scenarios that we were not able to collect in realistic conditions. The track 1 test clips can be found in the track_1 sub-directory of ICASSP_dev_test_set.
4.1.1 Real recordings
The real recordings consist of non-English and English segments. The English segment will have 300 clips that are from the blind test set from DNS Challenge 1. These clips were collected using the crowdsourcing platform and internally at Microsoft using a variety of devices, acoustic and noisy conditions. The non-English segment comprises of 100 clips in the following languages: Non-tonal: Portuguese, Russian, Spanish, and Tonal: Mandarin, Cantonese, Punjabi, and Vietnamese. In total, there are 400 real test clips.
4.1.2 Synthetic test set
The synthetic test set consists of 200 noisy clips obtained by mixing clean speech (non-English, emotional speech, and singing) with noise. The test set noises is taken from Audioset and Freesound [8]. The 100 non-English test clips include German, French and Italian languages from Librivox audio books. Emotion clean speech consists of laughter, yelling, and crying chosen from Freesound and mixed with test set noise to generate 50 noisy clips. Similarly, clean singing voice from Freesound was used to generate 50 noisy clips for the singing test set.
4.1.3 Blind test set
The blind test set for track 1 contains 700 noisy speech clips out of which 650 are real recordings and 50 synthetic noisy singing clips. It contains the following categories: (i) emotional (102 clips), (ii) English (276 clips), (iii) non-English including tonal (272 clips), (iv) tonal languages (112 clips) and (v) singing (50 clips). The real recordings were collected using the crowdsourcing platform and internally at Microsoft. This is the most diverse publicly available test set for a noise suppression task.
4.2 Track 2
For the pDNS track, we provide 2 minutes of clean adaptation data for each primary speaker with the goal to suppress neighboring speakers and background noise. pDNS models are expected to leverage speaker-aware training and speaker-adapted inference. There are two motivations to provide clean speech for the primary speaker: (1) speaker models are sensitive to false-alarms in speech activity detection (SAD) [24]; clean speech can be used for obtaining accurate SAD labels. (2) speaker adaptation is expected to work well using multi-conditioned data; clean speech can be used for generating reverberant and noisy data for speaker adaptation.
4.2.1 Real Recordings
The development test set contains 100 real recordings from 20 primary speakers collected using crowdsourcing. Each primary speaker has noisy test clips for three scenarios: (i) primary speaker in presence of neighboring speaker; (ii) primary speaker in presence of background noise; and (iii) primary speaker in presence of both background noise and neighboring speaker.
4.2.2 Synthetic test clips
The synthetic clips include 500 noisy clips from 100 primary speakers. Each primary speaker has 2 minutes of clean adaptation data. All clips have varying levels of neighboring speakers and noise. test set noise from Track 1 was mixed with primary speech extracted from VCTK corpus [25]. We used VoxCeleb2 [26] corpus for neighboring speakers.
4.2.3 Blind test set
The blind test set for track 2 contains 500 noisy speech real recordings from 80 unique speakers. All the real recordings were collected using the crowdsourcing platform. The noise source in the majority of these clips is a secondary speaker. We provided 2 mins clean speech utterances for each of the primary speakers that could be used to adapt the noise suppressor. All the utterances were in English.
5 Challenge Results
5.1 Evaluation Methodology
Most DNS evaluations use objective measures such as and PESQ [27], SDR, and POLQA [28]. However, these metrics are shown to not correlate well with subjective speech quality in the presence of background noise [29]. Subjective evaluation is the gold standard for this task. Hence, the final evaluation was done on the blind test set using the crowdsourced subjective speech quality metric based on ITU P.808 [9] to determine the DNS quality. For track 2, we appended each processed clip with a 5 secs utterance of the primary speaker at the beginning and 1 sec silence. We modified P.808 slightly to instruct the raters to focus on the quality of the voice of the primary speaker in the remainder of the processed segment. We conducted a reproducibility test with this change to P.808 and found that the average Spearman Rank Correlation Coefficient (SRCC) between the 5 runs was 0.98. Hence, we concluded that the change is valid. We used 5 raters per clip. This resulted in a 95% confidence interval (CI) of 0.03. We also provided the baseline noise suppressor [30] for the participants to benchmark their methods.
![[Uncaptioned image]](https://cdn.awesomepapers.org/papers/d3ad8a64-f81c-4e94-93ae-211aa659cb76/x1.png)
5.2 Evaluation Results
5.2.1 Track 1
A total of 16 teams from academia and industry participated in track 1. Table 1 shows the categorized P.808 results for track 1. The submissions are stack ranked based on the overall Differential MOS (DMOS). DMOS is the difference in MOS between the processed set and the original noisy. We can observe from the results that most of methods struggled to do well on singing and emotional clips. The noise suppressors tend to suppress the certain emotional and singing segments. Overall, the results show that performance of noise suppressors are not great in singing and emotional categories. The results also shows that the training data must balanced between English, non-English and tonal languages for the model to generalize well. It is important to include singing and other emotions as the ground truth to achieve better quality in these categories. Only half of the submissions did better than the baseline. The best model is only 0.53 DMOS better than the noisy, which is of absolute MOS 3.38. This shows that we are far away from achieving a noise suppressor that works robustly in almost all conditions.
5.2.2 Track 2
This track is the first of its kind and there is not much work done in this field. There were only 2 teams who participated in track 2. Each team submitted one set of processed clips with using speaker information for model adaptation and the other set without explicitly using speaker information. The results are shown in table 2. The best model gave only 0.14 DMOS. This shows that the problem of using speaker information to adapt the model is still in the infancy stage.
![[Uncaptioned image]](https://cdn.awesomepapers.org/papers/d3ad8a64-f81c-4e94-93ae-211aa659cb76/x2.png)
6 Summary & Conclusions
The ICASSP 2021 DNS Challenge was designed to advance the field of real-time noise suppression optimized for human perception in challenging noisy conditions. Large inclusive and diverse training and test datasets with supporting scripts were open sourced. Many participants from both industry and academia found the datasets very useful and submitted their enhanced clips for final evaluation. Only two teams participated in the personalized DNS track, which also shows that the field is in its nascent phase.
References
- [1] Y. Ephraim and D. Malah, “Speech enhancement using a minimum-mean square error short-time spectral amplitude estimator,” IEEE TASLP, 1984.
- [2] C. Karadagur Ananda Reddy, N. Shankar, G. Shreedhar Bhat, R. Charan, and I. Panahi, “An individualized super-gaussian single microphone speech enhancement for hearing aid users with smartphone as an assistive device,” IEEE Signal Processing Letters, vol. 24, no. 11, pp. 1601–1605, 2017.
- [3] S. Fu, Y. Tsao, X. Lu, and H. Kawai, “Raw waveform-based speech enhancement by fully convolutional networks,” in 2017 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA ASC).
- [4] Hyeong-Seok Choi, Hoon Heo, Jie Hwan Lee, and Kyogu Lee, “Phase-aware single-stage speech denoising and dereverberation with U-net,” arXiv preprint arXiv:2006.00687, 2020.
- [5] Yuichiro Koyama, Tyler Vuong, Stefan Uhlich, and Bhiksha Raj, “Exploring the best loss function for DNN-based low-latency speech enhancement with temporal convolutional networks,” arXiv preprint arXiv:2005.11611, 2020.
- [6] Jean-Marc Valin et al., “A perceptually-motivated approach for low-complexity, real-time enhancement of fullband speech,” arXiv preprint arXiv:2008.04259, 2020.
- [7] Umut Isik et al., “PoCoNet: Better speech enhancement with frequency-positional embeddings, semi-supervised conversational data, and biased loss,” arXiv preprint arXiv:2008.04470, 2020.
- [8] Chandan KA Reddy et al., “The INTERSPEECH 2020 deep noise suppression challenge: Datasets, subjective testing framework, and challenge results,” in ISCA INTERSPEECH, 2020.
- [9] Babak Naderi and Ross Cutler, “An open source implementation of ITU-T recommendation P.808 with validation,” in ISCA INTERSPEECH, 2020.
- [10] Chandan KA Reddy et al., “A scalable noisy speech dataset and online subjective test framework,” arXiv preprint arXiv:1909.08050, 2019.
- [11] V. Panayotov, G. Chen, D. Povey, and S. Khudanpur, “Librispeech: An ASR corpus based on public domain audio books,” in IEEE ICASSP, 2015.
- [12] Julia Wilkins, Prem Seetharaman, Alison Wahl, and Bryan Pardo, “Vocalset: A singing voice dataset.,” in ISMIR, 2018.
- [13] Houwei Cao, David G Cooper, Michael K Keutmann, Ruben C Gur, Ani Nenkova, and Ragini Verma, “CREMA-D: Crowd-sourced emotional multimodal actors dataset,” IEEE Trans. on Affective Computing, vol. 5, no. 4, pp. 377–390, 2014.
- [14] Zhiyong Zhang Dong Wang, Xuewei Zhang, “THCHS-30 : A free chinese speech corpus,” 2015.
- [15] Hui Bu, Jiayu Du, Xingyu Na, Bengu Wu, and Hao Zheng, “Aishell-1: An open-source mandarin speech corpus and a speech recognition baseline,” in 2017 20th Conference of the Oriental Chapter of the International Coordinating Committee on Speech Databases and Speech I/O Systems and Assessment (O-COCOSDA). IEEE.
- [16] “The Spoken Wikipedia Corpora,” https://nats.gitlab.io/swc/, [Online; accessed 2020-09-01].
- [17] “Telecooperation German Corpus for Kinect,” http://www.repository.voxforge1.org/downloads/de/german-speechdata-TUDa-2015.tar.gz, [Online; accessed 2020-09-01].
- [18] “M-AILABS Speech Multi-lingual Dataset,” https://www.caito.de/2019/01/the-m-ailabs-speech-dataset/, [Online; accessed 2020-09-01].
- [19] J. F. Gemmeke, D. P. W. Ellis, D. Freedman, A. Jansen, W. Lawrence, R. C. Moore, M. Plakal, and M. Ritter, “Audio set: An ontology and human-labeled dataset for audio events,” in IEEE ICASSP, 2017.
- [20] Joachim Thiemann, Nobutaka Ito, and Emmanuel Vincent, “The diverse environments multi-channel acoustic noise database (demand): A database of multichannel environmental noise recordings,” The Journal of the Acoustical Society of America, p. 3591, 05 2013.
- [21] Tom Ko, Vijayaditya Peddinti, Daniel Povey, Michael L Seltzer, and Sanjeev Khudanpur, “A study on data augmentation of reverberant speech for robust speech recognition,” in IEEE ICASSP, 2017.
- [22] Poju Antsalo et al., “Estimation of modal decay parameters from noisy response measurements,” in Audio Engineering Society Convention 110, 2001.
- [23] Hannes Gamper, “Blind C50 estimation from single-channel speech using a convolutional neural network,” in Proc. IEEE MMSP, 2020, pp. 136–140.
- [24] John HL Hansen and Taufiq Hasan, “Speaker recognition by machines and humans: A tutorial review,” IEEE Signal processing magazine, vol. 32, no. 6, pp. 74–99, 2015.
- [25] Junichi Yamagishi, Christophe Veaux, Kirsten MacDonald, et al., “CSTR VCTK corpus: English multi-speaker corpus for CSTR voice cloning toolkit (version 0.92),” 2019.
- [26] Joon Son Chung, Arsha Nagrani, and Andrew Zisserman, “Voxceleb2: Deep speaker recognition,” ISCA INTERSPEECH, 2018.
- [27] “ITU-T recommendation P.862: Perceptual evaluation of speech quality (PESQ): An objective method for end-to-end speech quality assessment of narrow-band telephone networks and speech codecs,” Feb 2001.
- [28] John Beerends et al., “Perceptual objective listening quality assessment (POLQA), the third generation ITU-T standard for end-to-end speech quality measurement part II-perceptual model,” AES: Journal of the Audio Engineering Society, vol. 61, pp. 385–402, 06 2013.
- [29] A. R. Avila, H. Gamper, C. Reddy, R. Cutler, I. Tashev, and J. Gehrke, “Non-intrusive speech quality assessment using neural networks,” in IEEE ICASSP, 2019.
- [30] Sebastian Braun and Ivan Tashev, “Data augmentation and loss normalization for deep noise suppression,” 2020.