This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Speakerfilter-Pro: an improved target speaker extractor combines the time domain and frequency domain

Abstract

This paper introduces an improved target speaker extractor, referred to as Speakerfilter-Pro, based on our previous Speakerfilter model. The Speakerfilter uses a bi-direction gated recurrent unit (BGRU) module to characterize the target speaker from anchor speech and use a convolutional recurrent network (CRN) module to separate the target speech from a noisy signal. Different from the Speakerfilter, the Speakerfilter-Pro sticks a WaveUNet module in the beginning and the ending, respectively. The WaveUNet has been proven to have a better ability to perform speech separation in the time domain. In order to extract the target speaker information better, the complex spectrum instead of the magnitude spectrum is utilized as the input feature for the CRN module. Experiments are conducted on the two-speaker dataset (WSJ0-mix2) which is widely used for speaker extraction. The systematic evaluation shows that the Speakerfilter-Pro outperforms the Speakerfilter and other baselines, and achieves a signal-to-distortion ratio (SDR) of 14.95 dB.

Index Terms—  Speaker extraction, Speakerfilter-Pro, neural network, anchor information

1 Introduction

The human brain can focus auditory attention on specific sounds by shielding the sound background in the presence of multiple speakers and background noise. This is the so-called cocktail party problem [1]. Recently, deep learning-based speech separation approaches for solving the cocktail party problem have attracted a considerable attention. Some of the representative advances include deep clustering [2], deep attractor network [3], and permutation invariant training [4]. More recently, a convolutional time-domain audio separation network (Conv-TasNet) [5] has been proposed and obtained a great separation performance improvement.

Lots of researchers separate all signals from the mixture signal to deal with the cocktail problem. However, in some practical situations, such as smart personal devices, we may be interested in recovering a target speaker while reducing background noise and other interfering speakers, called target speaker extraction.

Speech extraction exploits an anchor clue to identify and extracts the target speaker from a mixture signal. Recently, research on speaker extraction has gained increasing attention, as it naturally avoids the global permutation ambiguity issue and does not require knowing the number of sources in the mixtures. Several different frameworks based on frequency domain have been developed in recent years, including TENet [6], SpeakerBeam [7], VoiceFilter [8] and Speakerfilter [9]. There are also some successful frameworks in the time domain, such as TD-speakerbeam [10] for multi-channel speaker extraction.

Inspired by the better performances obtained in the time domain than frequency domain in the audio separation task, we propose an improved target speaker extractor, referred to as Speakerfilter-Pro, based on our previous Speakerfilter [9] model. Different from the Speakerfilter, the Speakerfilter-Pro adds a WaveUNet [11] module in the beginning and the ending, respectively. The WaveUNet is based on the time domain has been proven that it has a better ability in speech separation. We still use the convolutional recurrent network (GCRN) to characterize the target speaker information. To extract the target speech better, the complex spectrum instead of the magnitude spectrum is used as the input feature for the GCRN module.

Compared with our previous work, there has two contributions. 1) The WaveUNet is used to separate the target speech better. 2) The complex spectrum feature is employed to make better use of the anchor speech. Through the experiments, we find that the proposed method outperforms Speakerfilter and other baselines, and we also find that WaveUNet and the complex spectrum feature can boost each other.

In this paper, we first explain how this work relates to our previous work (Section 2). Then, we describe the proposed Speakerfilter-Pro (Section 3). Section 4 describes the experimental setup and result. We conclude this paper in Section 5.

2 Previous work

2.1 Problem formulation

The purpose of speaker extraction is to extract the target speech from an observed mixture signal. The work takes anchor speech as auxiliary information. The mixture signal can be written as:

y=st+si,y=s_{t}+s_{i}, (1)

where sts_{t} and sis_{i} are the target speech signal and the interfering signal, respectively. The interfering signal may comprise the speech signal from other speakers and background noise. In this paper we only consider the interfering speech. The speaker extractor can be defined as:

s^t=f(y,sa),\hat{s}_{t}=f(y,s_{a}), (2)

where ff is the transformation carried out by Speakerfilter-Pro, s^t\hat{s}_{t} is the estimated target speech, and sas_{a} is the anchor speech.

2.2 Speakerfilter system

The proposed method is based on our previous work [9], and here we give a brief introduction of the Speakerfilter. The Speakerfilter system takes two speech signals as input: noisy speech and anchor speech. A bi-direction gated recurrent unit (BGRU) is used to encode the anchor speech and average the outputs of BGRU in the time dimension to produce the anchor-vector. This anchor-vector can be viewed as the identity vector of the target speaker. Then, it is stacked to each frame of the mixture magnitude spectrum. We use a gated convolutional network (GCNN) to encode the anchor vector into each layer of the speech separator built by a GCRN. The GCRN is used to extract the target speech from noisy speech, with the help of the anchor-vector.

3 Speakerfilter-Pro

3.1 WaveUNet

Recently, performs speech separation in the time domain is very attractive. Researchers are also proven that the speech separation performs better in the time domain than in the frequency domain [10].

The WaveUNet [11] is the most commonly used model in the time-domain. The structure is shown in Fig. 1. It computes an increasing number of higher-level features on coarser time scales using downsampling (DS) blocks. These features are combined with the earlier computed local, high-resolution features using upsampling (US) blocks, yielding multi-scale features to make predictions.

Refer to caption
Fig. 1: WaveUNet structure

3.2 Model topology

Inspired by the WaveUNet’s powerful speech separation capabilities, we propose an improved target speaker extractor, referred to as Speakerfilter-Pro, based on our previous work [9]. Different from the Speakerfilter, the Speakerfilter-Pro add a WaveUNet module in the beginning and the ending, respectively. Second, in order to encode the target speaker better, the complex spectrum instead of the magnitude spectrum is used as the input feature for the GCRN module.

Refer to caption
Fig. 2: Overview of the Speakerfilter-Pro system.

The structure of the Speakerfilter-Pro is shown in Fig. 2. First, the time-domain mixture signal is fed into the beginning WaveUNet. After obtaining the output of the WaveUNet, the short-time Fourier transform (STFT) is used to transfer the time domain signal to the frequency domain. The real and the image part of the spectrum (RI) then stack with the anchor-vector as the input to the GCRN, where the target speaker-vector is obtained by the speaker feature extractor. With an encoder-decoder architecture, the CRN encodes the input features into a higher-dimensional latent space and then models the latent feature vectors’ sequence via two bi-direction long short-term memory (Bi-LSTM) layers. The output sequence of the Bi-LSTM layers is subsequently converted back to the original input size by the decoder. To improve the flow of information and gradients throughout the network, skip connections are used to concatenate each encoder layer’s output to the input of the corresponding decoder layer. After GCRN, the Deepfilter [12] is utilized to recover the target speech better. Deepfilter obtains enhanced speech X^\hat{X} from mixture speech XX by applying a complex filter:

X^(n,k)=i=IIl=LLHn,k(l+L,i+I)X(nl,ki)\hat{X}(n,k)=\sum_{i=-I}^{I}\sum_{l=-L}^{L}H^{*}_{n,k}(l+L,i+I)\cdot X(n-l,k-i) (3)

where 2L+12\cdot L+1 is the filter dimension in time-frame direction, 2I+12\cdot I+1 in frequency direction and HH^{*} is the complex conjugated 2D filter of TF bin (n, k). In our experiments, we set LL = 2 and II = 1 resulting in a filter dimension of (5, 3). Finally, the WaveUNet is used in the last part to improve the performance further.

For speaker feature extractor, the magnitude spectrum of anchor speech in frequency-domain is used as the inputs. We firstly encode the anchor speech into 129-dimensional features by using BGRU and average the outputs of BGRU in the time dimension. This 129-dimensional anchor vector can be viewed as the identity vector of the target speaker. Then, it is stacked to each frame of the mixture feature. To utilize the speaker information effectively, we propose to use CNN to encode the anchor vector into each layer of CRN, as shown in Fig. 2.

The whole modules are training jointly. The training loss is defined in the frequency domain. The loss function is mean squared error (MSE), which is given by:

Lmse=1NKn=1Nk=1KX(n,k)X^(n,k)2L_{mse}=\frac{1}{N*K}\sum_{n=1}^{N}\sum_{k=1}^{K}\|X(n,k)-\hat{X}(n,k)\|^{2} (4)

where NN is the number of TF unit, XX and X^\hat{X} represent the short-time Fourier transform (STFT) of clean and estimated target speech, respectively.

4 EXPERIMENTS

4.1 Dataset

The proposed system is evaluated by the WSJ0-2mix datasets111Available at: http://www.merl.com/demos/deep-clustering. In WSJ0-2mix, each sentence contains two speakers. The WSJ0-2mix dataset introduced in [2] is derived from the WSJ0 corpus [13]. The 30h training set and the 10h validation set contain two-speaker mixtures generated by randomly selecting from 49 male and 51 female speakers from si_tr_s. The Signal-to-Noise Ratios (SNRs) are uniformly chosen between 0 dB and 5 dB. The 5h test set is generated similarly by using utterances from 16 speakers from si_et_05, which do not appear in the training and validation sets. The test set includes 1603 F&M sentences, 867 M&M sentences, and 530 F&F sentences.

For each mixture, we randomly choose an anchor utterance from the target speaker (different from the utterance in the mixture).

4.2 Baseline model and metrics setting

We compare the proposed algorithm with two baselines: SpeakerBeam+DC[7] and Speakerfilter [9].

Compared with the Speakerfilter, the proposed method has two improvements: 1) the complex spectrum, not the magnitude spectrum, is used in the GCNN part, which can characterize the target speaker better. 2), The time-domain WaveUNet is used in the beginning and ending, which can separate the target speaker better. The proposed method also can be referred to as Speakerfilter+RI+WaveUNet. Based on this, we set up two comparison methods to illustrate the improvement’s impact on the separation effect: Speakerfilter+RI and Speakerfilter+WaveUNet.

The performance is evaluated with two objective metrics: perceptual evaluation of speech quality (PESQ) [14] and Signal-to-Distortion Ratio (SDR) [15]. For both of the PESQ and SDR metrics, the higher number indicates better performance.

4.3 Evaluation results

Table 1: Average SDR score on the test set for different methods.
Method SDR
unprocessed 2.62
SpeakerBeam+DC [7] 13.50
Speakerfilter [9] 13.89
Proposed 14.95

Table 1 presents the comprehensive evaluation for different approaches on SDR. It can be seen that the proposed method obtains the best performance. The baseline models are using the T-F domain only for target speaker extraction. However, the proposed method leverages the time-domain modeling ability of the WaveUNet and the frequency-domain modeling ability of the GCRN, simultaneously, that can obtain a better performance. Compared with the unprocessed, the proposed method improves the SDR nearly 12.33 dB.

Table 2: The SDR performance under different modifications on Speakerfilter.
F&M F&F M&M Average
unprocessed 2.58 2.71 2.65 2.62
Speakerfilter 15.80 11.80 11.64 13.89
Speakerfilter+RI 15.96 11.72 11.78 13.86
Speakerfilter+WaveUNet 16.16 12.84 12.67 14.26
Speakerfilter+RI+WaveUNet 16.52 12.91 13.68 14.95
Table 3: The PESQ performance under different modifications on Speakerfilter.
F&M F&F M&M Average
unprocessed 2.15 2.13 2.25 2.17
Speakerfilter 3.30 2.90 2.91 3.12
Speakerfilter+RI 3.26 2.87 2.91 3.09
Speakerfilter+WaveUNet 3.32 3.01 2.97 3.17
Speakerfilter+RI+WaveUNet 3.37 3.01 3.12 3.24

As shown in Table 2 and 3, most of the modifications have positive effect on the baseline. On average, the Speaker+RI+WaveUNet model obtains the best results.

The WavUNet has a large improvement for Speakerfilter and Speakerfilter+RI models. The average SDR improves 0.37 (from 13.89 to 14.26) and 1.09 (from 13.86 to 14.95) dB, respectively. The average PESQ improves 0.05 (from 3.12 to 3.17) and 0.15 (from 3.09 to 3.24) dB, respectively. We should note that, after adding the WaveUNet module, the RI-based model (SDR=1.09,PESQ=0.15\triangle SDR=1.09,\triangle PESQ=0.15) is more effective than the amplitude spectrum-based model (SDR=0.37,PESQ=0.05\triangle SDR=0.37,\triangle PESQ=0.05).

For the Speakerfilter system, after changing the magnitude spectrum to the complex spectrum, both the SDR and the PESQ are reduced by 0.03 dB. However, in the Speakerfilter+WaveUNet system, after changing the magnitude spectrum to the complex spectrum, the SDR and the PESQ are increased by 0.69 and 0.07 dB, respectively, which means that the complex spectrum feature can boost the speaker extractor performance with the participation of WaveUNet.

As mentioned above, the complex spectrum feature and the WaveUNet can boost each other.

Refer to caption
(a) Speakerfilter result
Refer to caption
(b) Speakerfilter-Pro result
Fig. 3: Spectrograms of extracted speech using Speakerfilter and Speakerfilter-Pro. (a) The spectrogram of the separated target speech using Speakerfilter. (b) The spectrogram of the separated target speech using the Spekaerfilter-Pro.

Fig. 3 illustrates the spectrogram of separated speech using Speakerfilter and Speakerfilter-Pro, respectively. Our method has better performance on target speaker separation. As shown in the yellow rectangle, the proposed method can suppress the interfering speaker better than Speakerfilter.

5 Conclusions

This paper proposes a target speaker extractor, where WaveUNet, CRN, and BGRU-GCNN are used for time-domain feature enhancer, speech separator, and speaker feature extractor, respectively. By leveraging the time-domain modeling ability of WaveUNet and the frequency-domain modeling ability of the CRN, the target speaker can be tracked well. According to the experiment results, the proposed method achieves better performance than other baseline methods. In the future, we will explore the robustness problem in the presence of noise and reverberation.

6 Acknowledgments

This research was partly supported by the China National Nature Science Foundation (No. 61876214).

References

  • [1] Stephan Getzmann, Julian Jasny, and Michael Falkenstein, “Switching of auditory attention in ’cocktail-party’ listening: Erp evidence of cueing effects in younger and older adults,” Brain and cognition, vol. 111, pp. 1–12, 2017.
  • [2] John R Hershey, Zhuo Chen, Jonathan Le Roux, and Shinji Watanabe, “Deep clustering: Discriminative embeddings for segmentation and separation,” in 2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, 2016, pp. 31–35.
  • [3] Zhuo Chen, Yi Luo, and Nima Mesgarani, “Deep attractor network for single-microphone speaker separation,” in 2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, 2017, pp. 246–250.
  • [4] Dong Yu, Morten Kolbæk, Zheng-Hua Tan, and Jesper Jensen, “Permutation invariant training of deep models for speaker-independent multi-talker speech separation,” in 2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, 2017, pp. 241–245.
  • [5] Yi Luo and Nima Mesgarani, “Conv-tasnet: Surpassing ideal time–frequency magnitude masking for speech separation,” IEEE/ACM transactions on audio, speech, and language processing, vol. 27, no. 8, pp. 1256–1266, 2019.
  • [6] Wenjie Li, Pengyuan Zhang, and Yonghong Yan, “Tenet: target speaker extraction network with accumulated speaker embedding for automatic speech recognition,” Electronics Letters, 2019.
  • [7] Kateřina Žmolíková, Marc Delcroix, Keisuke Kinoshita, Tsubasa Ochiai, Tomohiro Nakatani, Lukáš Burget, and Jan Černockỳ, “Speakerbeam: Speaker aware neural network for target speaker extraction in speech mixtures,” IEEE Journal of Selected Topics in Signal Processing, vol. 13, no. 4, pp. 800–814, 2019.
  • [8] Quan Wang, Hannah Muckenhirn, Kevin Wilson, Prashant Sridhar, Zelin Wu, John Hershey, Rif A Saurous, Ron J Weiss, Ye Jia, and Ignacio Lopez Moreno, “Voicefilter: Targeted voice separation by speaker-conditioned spectrogram masking,” arXiv preprint arXiv:1810.04826, 2018.
  • [9] Shulin He, Hao Li, and Xueliang Zhang, “Speakerfilter: Deep learning-based target speaker extraction using anchor speech,” in ICASSP 2020-2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, 2020, pp. 376–380.
  • [10] Marc Delcroix, Tsubasa Ochiai, Katerina Zmolikova, Keisuke Kinoshita, Naohiro Tawara, Tomohiro Nakatani, and Shoko Araki, “Improving speaker discrimination of target speech extraction with time-domain speakerbeam,” in ICASSP 2020-2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, 2020, pp. 691–695.
  • [11] Daniel Stoller, Sebastian Ewert, and Simon Dixon, “Wave-u-net: A multi-scale neural network for end-to-end audio source separation,” arXiv preprint arXiv:1806.03185, 2018.
  • [12] Wolfgang Mack and Emanuël AP Habets, “Deep filtering: Signal extraction using complex time-frequency filters,” arXiv preprint arXiv:1904.08369, 2019.
  • [13] John Garofalo, David Graff, Doug Paul, and David Pallett, “Csr-i (wsj0) complete,” Linguistic Data Consortium, Philadelphia, 2007.
  • [14] Antony W Rix, John G Beerends, Michael P Hollier, and Andries P Hekstra, “Perceptual evaluation of speech quality (pesq)-a new method for speech quality assessment of telephone networks and codecs,” in 2001 IEEE International Conference on Acoustics, Speech, and Signal Processing. Proceedings (Cat. No. 01CH37221). IEEE, 2001, vol. 2, pp. 749–752.
  • [15] Emmanuel Vincent, Rémi Gribonval, and Cédric Févotte, “Performance measurement in blind audio source separation,” IEEE transactions on audio, speech, and language processing, vol. 14, no. 4, pp. 1462–1469, 2006.