LIMUSE: LIGHTWEIGHT MULTI-MODAL SPEAKER EXTRACTION
Abstract
Multi-modal cues, including spatial information, facial expression and voiceprint, are introduced to the speech separation and speaker extraction tasks to serve as complementary information to achieve better performance. However, the introduction of these cues brings about an increasing number of parameters and model complexity, which makes it harder to deploy these models on resource-constrained devices. In this paper, we alleviate the aforementioned problem by proposing a Lightweight Multi-modal framework for Speaker Extraction (LiMuSE). We propose to use GC-equipped TCN, which incorporates Group Communication (GC) and Temporal Convolutional Network (TCN) in the Context Codec module, the audio block and the fusion block. The experiments on the MC_GRID dataset demonstrate that LiMuSE achieves on par or better performance with a much smaller number of parameters and less model complexity. We further investigate the impacts of the quantization of LiMuSE. Our code and dataset are provided111https://github.com/aispeech-lab/LiMuSE.
Index Terms: Model Compression, Speaker Extraction, Multi-Modality, Group Communication, Quantization
(0cm,11cm) Copyright 2022 IEEE. Published in the 2022 IEEE Spoken Language Technology Workshop (SLT) (SLT 2022), scheduled for 19-22 January 2023 in Doha, Qatar. Personal use of this material is permitted. However, permission to reprint/republish this material for advertising or promotional purposes or for creating new collective works for resale or redistribution to servers or lists, or to reuse any copyrighted component of this work in other works, must be obtained from the IEEE. Contact: Manager, Copyrights and Permissions / IEEE Service Center / 445 Hoes Lane / P.O. Box 1331 / Piscataway, NJ 08855-1331, USA. Telephone: + Intl. 908-562-3966.
1 Introduction
The Cocktail Party Problem is a fundamental topic in auditory scene analysis, which refers to the phenomenon that humans are able to focus his/her attention on specific auditory stimuli by masking out the acoustic background noise in complex auditory scenes [1]. With the rapid development of deep learning technologies, there is an emerging trend of using deep-learning-based methods to model the Cocktail Party Problem. In the speech processing community, these approaches can be categorized into two types: blind speech separation and target speaker extraction. Blind speech separation aims to mimic the human’s bottom-up stimulus-driven attention by simultaneously estimating the speech of different speakers from the mixture, which is often confronted with two problems: permutation problem [2] and output dimension mismatch problem [3]. By contrast, speaker extraction (also termed as target speech separation) only extracts the speech of interest from the overlapped audio of simultaneous speakers. It emulates humans’ top-down voluntary focus [4] by utilizing the information from the target speaker as prior knowledge to direct the attention to the interested speaker. In this way, speaker extraction avoids the permutation problem and output mismatch problem inherently.
Speaker extraction often requires additional information that identifies the target speaker. Voiceprint, which characterizes a speaker, has proved to be an important cue in auditory scene analysis [5, 4, 6, 7]. Recently, incorporating visual information such as face or lip image sequences, in which phonetic or prosaic content coincides with muscle movements, into speaker extraction systems becomes an emerging research focus to improve the robustness and separation quality [8, 9, 10, 11]. What’s more, multi-channel-based methods can benefit from additional direction information to further improve the performance [12, 13]. By utilizing all the available information of the target speaker mentioned above, a multi-channel multi-modal system is proposed [14] and achieves superb performance. In challenging acoustic environments, the acoustic information of the target speaker can be blurry. The information of other modalities serves as complementary and steady information to increase the quality of the interested speech. However, the addition of multi-modal information, particularly visual information, typically results in larger model sizes and higher computational costs, which poses severe challenges for the deployment of these deep-learning-based models on resource-constrained devices, such as mobile phones and other edge devices.
In order to alleviate the deployment problems on resource-constrained devices, there are mainly two types of methods: design of novel model architectures and model compression methods. MobileNet [15] proposes to use depth-wise separable convolution to decrease the parameters of the traditional convolutional layer. MobileNetV2 [16] further proposes linear bottleneck and inverted residual for lightweight model structure design. The practice of Group Communication [17] proves that a small, group-shared module is adequate to preserve the model capacity given the inter-group modeling step. Furthermore, Group Communication with Context Codec (GC3) [18] cultivates a more efficient model design to decrease model size without sacrificing performance. Besides, various model compression methods have been proposed to lighten deep neural networks, such as parameter pruning [19, 20], knowledge distillation [21, 22] and quantization [23, 24, 25]. In contrast to the adjustments of model architectures, the quantization technology compresses a model by converting a full-precision neural network into a low-bitwidth integer version. By limiting the model topology and quantizing the model to 8-bit integer form, VoiceFilter-Lite [5] makes the model tiny and fast to meet strict on-device production requirements. TinyWASE [26] investigates the effects of quantization function in audio-only speaker extraction tasks and greatly compresses the model size without sacrificing much performance.

In this paper, we propose a lightweight multi-modal speaker extraction framework, which incorporates multi-channel information, target speaker’s visual feature and voiceprint as reference information and further apply Group Communication, Context Codec and an ultra-low bit quantization technology to reduce the model size and complexity while maintaining relatively high performance. Our contributions can be concluded as follows:
-
1.
As far as we know, it is the first time that Group Communication has been explored in a multi-modal speaker extraction model. The results have shown the effectiveness, and an appropriate number of groups may promote the inter-modal information exchange and even achieve improvements;
-
2.
We incorporate GC-equipped TCN blocks in the process of sequence modeling and Context Codec to decrease both the width of neural network and length of sequential feature without using RNN-related structures, which is more friendly to incorporate quantization technologies, as attempts for quantization of RNNs may show considerable performance degradation with low bit-width weights and activations;
-
3.
We further compress the model size with quantization technologies. The results have shown that our model with ultra-low bit quantization can also maintain a comparable performance, compared with the baselines.
2 Model Design
Our proposed model is a multi-stream architecture that takes multi-channel audio mixture, target speaker’s enrolled utterance and visual sequences of detected faces as inputs, and outputs the target speaker’s mask in the time domain. The encoded audio representations of the mixture are then multiplied by the generated mask to obtain the target audio. Figure 1 provides an overview of our network.
2.1 Audio and Video Representations
2.1.1 Audio Stream
The audio encoder encodes the multi-channel audio mixture into spectrum-like representation in the time domain, and the audio decoder is used to transform the masked encoder representation to the target waveform. The encoder and the decoder follow the same structure in [27], which uses a 1-dimensional convolutional layer and a 1-dimensional transpose convolutional layer, respectively.
Formulating the audio signals in the time domain has some advantages over the traditional time-frequency representations obtained by Short-Time Fourier Transform (STFT), including trainable weights, finer coding granularity and avoidance of phase reconstruction [7].
2.1.2 Speaker Embedding Stream
The voiceprint models the characteristics of the target speaker and serves as prior knowledge to help the model filter out the target speaker’s speech from the mixture. The voiceprint encoder takes the speaker embeddings of the target speaker as input. In this paper, we use an off-the-shelf speaker diarization toolkit pyannote [28] to extract the speaker embeddings , where denotes the length of the speaker embedding feature dimension.
In order to align with the time dimension of the mixed audio stream, we stack the speaker embeddings along the time dimension to form . The voiceprint encoder uses a fully connected (FC) layer to map the feature dimension of the voiceprint to obtain , which is the same shape as the representation of the audio mixture stream.
2.1.3 Video Stream
We obtain face embeddings of the target speaker from videos using the network in [29], which extracts speech-related visual features from visual inputs explicitly by the adversarially disentangled method. The face embedding extraction network first learns joint visual and auditory representation from clean audio and video pairs to establish the correlation between the audio features and the visual features, and then disentangles the speech-related visual features from the joint audio-visual representation with the adversarial training method.
The pretrained face embedding extraction network is trained on LRW dataset [30] and MS-Celeb-1M dataset [31] to obtain , where denotes the feature dimension of the face embedding and denotes the number of frames in each video stream. The visual encoder then maps the extracted face embeddings to match the feature dimension of the encoded audio features and obtain . Moreover, since the time resolution of the video stream and the audio stream is different, we upsample the face embeddings along the temporal dimension to synchronize the audio stream and the video stream by nearest neighbor interpolation to obtain . The face embeddings are then used to model the temporal synchronization and interaction between visemes and speech [32].
2.2 Separation Module
2.2.1 GC-Equipped TCN block
In order to reduce the number of parameters, we apply Group Communication (GC) [18] to the separation modules. Given a feature , it can be decomposed into groups of low-dimensional feature vectors . A GC module is applied to capture the inter-group dependencies.
We use the Transform-Average-Concatenate (TAC) module [34] for the GC module. It is similar to the squeeze-and-excitation design paradigm [35], where there is a residual connection between a pool of features and a global-averaged feature which is generated to summarize the input feature pool. For , a fully-connected layer followed by PReLU activation [36] is used for the transformation step to generate . Then all s are averaged and passed through another fully connected layer with PReLU for the average step to get . Finally, is concatenated with and passed to the third fully connected layer with PReLU to generate . A residual connection is added between and . It is demonstrated in [18] that using TAC for the Group Communication module achieves better performance and lower computational costs than other network structures, such as BLSTM and self-attention.
In this paper, Group Communication is applied before each Temporal Convolutional Network (TCN) block [27], which forms a GC-equipped TCN block, as is shown in Figure 1 (B). Each TCN block consists of a depth-wise separable convolution, PReLU activation [36] and a layer normalization operation. By introducing Group Communication to the TCN block, we can reduce the model width for the TCN block by (number of groups) times when there is no overlapping within the groups.
2.2.2 Context Codec with GC-Equipped TCN block
On one hand, speech has a rich temporal structure over multiple time scales [37]. On the other hand, shorter kernel length of the convolution in the audio encoder leads to better separation performance and higher model complexity at the same time. Motivated by the above two characteristics, Context Codec [18] is proposed to decrease sequence length, which is similar to a nonlinear downsampling step by compressing the context of feature vectors into a summary one and then decompress it back to the context. A Context Codec module typically contains a context encoder and a context decoder. In this paper, instead of using RNN [18], we propose to use the GC-equipped TCN network for the Context Codec module.
To be specific, given a sequence of feature , where denotes the number of feature channels, the context encoder first splits it along the temporal dimension into blocks with an overlap ratio of 50%, where denotes the number of context blocks and denotes the context size as illustrated in Figure 1 (C). Then the context encoder applies a GC-equipped TCN network (Section 2.2.1) on each to establish inter-group correlation and obtains . At last, a mean operation is applied on the context dimension of to produce a sequence of summarization vectors which represent the whole feature sequence. The compressed sequence can then be sent into feature processing module with . The context decoder adds the output of feature processing module to each time step in and applies another GC-equipped TCN network for the nonlinear transformation. Overlap-and-add is then applied on to form the sequence of feature of the original length . The implementation details can be found in our released source code.
2.2.3 Audio Block and Fusion Block
In order to capture the long-term dependencies in the encoded audio features, the audio block is composed of several GC-equipped TCN blocks. In the audio block, we use two repeats of the TCN networks, and each repeat of the TCN network is composed of several TCN blocks. The dilation factor of depth-wise convolution increases exponentially by two in each repeat, which enlarges the receptive field of the module to capture the correlation in a larger time scale.
The output of the audio block is gathered with the encoded voiceprint and visual features through concatenation. Then we use a repeat of GC-equipped TCN blocks with an exponential increasing dilation factor of depth-wise convolution in the fusion block to process the fused features. The voiceprint and visual cues serve as auxiliary information to help the model extract the target speech from the mixture. A convolutional layer is followed by the fusion block to generate the mask of the target speaker.
2.3 Ultra-low bit Quantization
In this section, we further investigate how to apply quantization strategies [26] for multi-modal scenarios.
2.3.1 Weight Quantization
For weight quantization, we propose to use the quantization function [38] to quantize weights to ultra-low bits for multi-modal speaker extraction tasks. The quantization function combines a linear combination of several simple functions to reformulate the ideal quantization operation and gradually approximates it by gradually increasing the temperature parameter during training.
Suppose that is the full-precision value to be quantized, is the quantized integer constrained to a predefined set . For example, if we perform 3-bit weight quantization, then can be defined as . The ideal quantization operation can be reformulated as a combination of unit step functions as follows:
(1) |
(2) |
where , is the scale factor of the input. Eq 2 defines the unit step function, where and is the bias for the unit step function. The global offset is defined as .
However, the unit step function in Eq 2 is non-differentiable and thus can not be trained from end to end. The quantization function replaces the unit step function with differentiable sigmoid function in the training procedure as follows:
(3) |
(4) |
where denotes the temperature parameter, and are learnable scale factors of the output and input, respectively. While in the inference phase, the ideal quantization function in Eq 1 is used. Performance may suffer if different quantization functions are used throughout the training and inference phases. To narrow the gap, temperature is introduced to the sigmoid function. When is large, the gap between the sigmoid function and the unit step function is minimal, and vice versa. Starting from a small temperature and gradually increasing it, the quantized neural networks can be well learned.
In this paper, we perform layer-wise non-uniform quantization, in which weights from different layers use different quantization functions. We quantize all neural network modules except the audio decoder. We also do not quantize the PReLU activation and the layer normalization function because they may be sensitive to quantization and have a small number of parameters. To perform non-uniform quantization, we use K-means clustering on full-precision weights to obtain ranked cluster centers in the ascending order. The bias is initialized as and gets fixed during training.
2.3.2 Activation Quantization
We also quantize the activations of the corresponding quantized layers to make matrix multiplication run faster. For activation quantization, we use the most commonly used min-max linear quantization [39], which is defined as:
(5) |
(6) |
where is the tensor to be quantized, is the element in , and are the minimum and maximum elements, respectively. In this paper, the activation quantization bits are , then in equation 6. Straight-Through Estimator (STE) [40] is used to backpropagate the gradients through quantized activations.
2.4 Loss Function
Scale-Invariant Signal-to-Distortion ratio (SI-SDR) between the predicted wave and the clean wave is used to optimize the network from end to end. The loss function is defined as:
(7) |
3 EXPERIMENT SETUP
3.1 Datasets
We evaluate our system on speaker extraction problems using the GRID dataset [41], which contains 18 male and 15 female speakers, and each of them has 1,000 frontal 3-second face video recordings. The sampling rate of speech is 16 kHz. And the videos are sampled at 25 FPS. We randomly select 3 males and 3 females to construct a validation set of 2.5 hours and another 3 males and 3 females for a test set of 2.5 hours. The rest of the speakers form the training set of 30 hours.
To construct a 2-speaker mixture, we randomly choose two different speakers and select audio from each chosen speaker, then mix these two clips at a random SNR level between -5 dB and 5 dB. Then we use the SMS-WSJ toolkit [42] to generate the simulated anechoic dual-channel audio mixture. Specifically, we place 2 microphones at the center of the room, and the distance between the microphones is 7 cm. The implementation code along with the Multi-Channel GRID (MC_GRID) dataset and the feature embeddings obtained from raw videos and reference speech used in our experiments are all available.
3.2 Training Details
We first train the full-precision network for 50 epochs with the initial learning rate set to . Then, to train the quantization network, we load the weights of the pretrained full-precision model to get a good starting point. In the quantization stage, the whole separation module and mask generation module are quantized. The weights of the encoders are also quantized. In particular, the weights and the activations are quantized to 3 bits and 8 bits, respectively. The temperature in the quantization function is set to 5 in the beginning and increases linearly with respect to the epochs, which is . The learning rate is halved if the accuracy on the validation set does not improve for 4 consecutive epochs. The training process would stop if the accuracy on the validation set does not improve for 6 consecutive epochs. Adam [43] is used as the optimizer. Gradient clipping with maximum L2-norm of 5 is applied during training.
The hyperparameters of the proposed neural network are shown in Table 1.
Symbol | Description | Value |
---|---|---|
Number of channels in audio encoder | 128 | |
Length of the filters (in audio samples) | 32 | |
Kernel size in convolutional blocks | 3 | |
Number of repeats in audio block | 2 | |
Number of repeats in fusion block | 1 | |
Context size (in frames) | 32 | |
Weight quantization bits | 3 | |
Activation quantization bits | 8 | |
Temperature increment per epoch | 5 |
Method | Task | Input Stream | SI-SDRi (dB) | SDRi (dB) | Model Size | MACs |
---|---|---|---|---|---|---|
LiMuSE | SE | A+Vp+Vis | 17.25 | 17.50 | 0.48MB | 7.52G |
MuSE [44]* | SE | A+Vis | 8.53 | - | 54.27MB | 25.87G |
AVMS [11] | SS | A+Vis | - | 15.74 | 22.34MB | 60.66G |
AVDC [9] | SS | A+Vis | - | 8.88 | - | - |
Conv-TasNet [27]** | SS | A | 14.98 | 15.27 | 13.28MB | 21.44G |
-
*
This paper only reports cross-dataset evaluation results on the GRID dataset. It is trained on VoxCeleb2 [45] and tested on GRID.
-
**
We re-implement and train Conv-Tasnet on GRID for comparison.
4 RESULTS AND DISCUSSIONS
4.1 LiMuSE vs baselines on GRID
We report SI-SDR improvement (SI-SDRi) and SDR improvement (SDRi) scores for the evaluation of the model performance. For model complexity, the model size and the number of MAC operations (MACs) are reported as metrics. MACs are calculated using an open-source toolbox222https://github.com/Lyken17/pytorch-OpCounter. As is shown in Table 2, LiMuSE exhibits competitive performance compared with other existing multi-modal speech separation and speaker extraction baselines with substantially smaller model size and model complexity.
Method | K | SI-SDR (dB) | SDR (dB) |
---|---|---|---|
Adversarially Disentangle | 32 | 21.75 | 22.61 |
16 | 24.27 | 24.83 | |
Deep AVSR [46] | 32 | 18.87 | 20.19 |
16 | 20.98 | 22.02 |
Method | K | SI-SDR (dB) | SDR (dB) | #Params |
---|---|---|---|---|
FC layer | 32 | 21.75 | 22.61 | 0.41M |
16 | 24.27 | 24.83 | 1.12M | |
V-TCN | 32 | 22.81 | 23.20 | 0.74M |
16 | 25.12 | 25.84 | 1.46M |
K | Causal | SI-SDR (dB) | SDR (dB) | SDRi (dB) |
---|---|---|---|---|
32 | 21.75 | 22.61 | 22.40 | |
✓ | 20.59 | 21.64 | 21.43 | |
16 | 24.27 | 24.83 | 24.63 | |
✓ | 22.23 | 22.93 | 22.72 |
Model | K | SI-SDR (dB) | SDR (dB) | SDRi (dB) | #Params | Model Size | MACs | |
---|---|---|---|---|---|---|---|---|
LiMuSE (vanilla) | - | 23.54 | 24.02 | 23.83 | 8.95M | 34.14MB (100%) | 53.34G (100%) | |
LiMuSE (GC) | 32 | 19.17 | 20.71 | 20.50 | 0.37M | 1.40MB (4.1%) | 5.94G (11.1%) | |
16 | 23.78 | 23.65 | 23.45 | 0.97M | 3.70MB (10.8%) | 11.77G (22.1%) | ||
LiMuSE (GC+CC) | 32 | 21.75 | 22.61 | 22.40 | 0.41M | 1.55MB (4.54%) | 3.98G (7.46%) | |
16 | 24.27 | 24.83 | 24.63 | 1.12M | 4.28MB (12.5%) | 7.52G (14.1%) | ||
LiMuSE (GC+CC+Q) | 32 | 15.53 | 16.67 | 16.46 | 0.41M | 0.19MB (0.56%) | 3.98G (7.46%) | |
16 | 17.25 | 17.71 | 17.50 | 1.12M | 0.48MB (1.41%) | 7.52G (14.1%) |
(bit) | K | SDR (dB) | Model Size (MB) |
---|---|---|---|
3 | 32 | 16.67 | 0.19 |
16 | 17.71 | 0.48 | |
4 | 32 | 18.30 | 0.24 |
16 | 18.86 | 0.61 |
Model | K | SI-SDR (dB) | SDR (dB) | SDRi (dB) |
---|---|---|---|---|
LiMuSE | 32 | 21.75 | 22.61 | 22.40 |
(GC+CC) | 16 | 24.27 | 24.83 | 24.63 |
LiMuSE | 32 | 19.17 | 20.71 | 20.50 |
(GC+CC-Vp) | 16 | 21.13 | 22.38 | 22.17 |
LiMuSE | 32 | 14.75 | 16.32 | 16.11 |
(GC+CC-Vis) | 16 | 18.57 | 20.75 | 20.54 |
4.2 Comparison with different visual feature processing modules
We compare our visual feature processing module to those used in current audio-visual speech separation or speaker extraction models. For the pretrained visual front-end, we use the entire face region as model inputs. As is shown in Table 3, the Adversarily Disentangle approach outperforms Deep AVSR [46] (includes a 3D convolutional layer followed by a 18-layer ResNet [47]) in terms of face embedding performance. For the visual encoder modules, we replace the widely adopted temporal convolutional block (V-TCN), which consists of 5 residual connected rectified linear unit (ReLU), batch normalization (BN) and depth-wise separable convolutional layers (DS Conv1D) [8, 44] with a simple fully connected (FC) layer. As is shown in Table 4, when K=32, compared to V-TCN, using FC layer achieves 55% less parameters while suffering a 1dB loss in SI-SDR and SDR. To achieve an extremely small model size, we use the FC layer for the visual encoder in this paper.
4.3 Performance under causal configurations
The target application of LiMuSE is for resource-constrained devices. Therefore, it’s necessary to discuss the online casual usage of the model. Here we explore the performance of LiMuSE under causal configurations. Table 5 indicates that the system still performs well under causal usage. The performance degradation could be the result of causal convolution and/or the cumulative layer normalization [27].
4.4 Ablation study of LiMuSE
We conducted an ablation study of LiMuSE to investigate the influence of lightweight designs and various modalities.
4.4.1 Effects of lightweight designs
We investigate the effects of the Group Communication, Context Codec and the ultra-low bit quantization. The results are listed in Table 6.
The incorporation of Group Communication (GC) and Context Codec (CC) allows LiMuSE to maintain on par or even better performance under a suitable number of groups with a much smaller model size and less model complexity. We can see from the table that LiMuSE (GC) slightly performs better than LiMuSE (vanilla) when with only 0.37M parameters. LiMuSE (GC+CC) with further improves the performance with 0.04M extra parameters but reduces the MACs from 11.77G to 7.52G compared to LiMuSE (GC). It suggests that an appropriate number of groups may promote the inter-modal information exchange and achieve considerable improvements. When , the number of parameters and the MACs of LiMuSE (GC) and LiMuSE (GC+CC) are further reduced, and the performance degradation of LiMuSE (GC+CC) is very limited compared to LiMuSE (vanilla). Intuitively, we can empirically conclude that the larger the value of K, the fewer the model parameters with some performance degradation in most cases. A trade-off can be considered between the model complexity and performance according to the needs of production.
Furthermore, we apply ultra-low bit quantization to compress the model size and achieves as much as 180 of compression ratio when . In an extreme case, the model size of LiMuSE is compressed to 0.19MB. Combing the results of Table 2, the quantized LiMuSE still achieves competitive results over the audio-only baseline model TasNet [27] and audio-visual baseline models AVMS [11] and AVDC [9]. Though in [26], the authors have shown that the ultra-low bit quantization technology can achieve on-par performance on audio-only TCN-based speaker extraction models, applying the ultra-low quantization to our audio-visual TCN-based LiMuSE model leads to some performance degradation when the weights are quantized to 3 bits. While deploying the model on resource-constrained devices, it is recommended that an appropriate quantization bit be chosen according to the trade-off between the performance and model size as is shown in Table 7.
4.4.2 Effects of the multimodal cues
Based on the experimental results in Section 4.4.1, we investigate the effects of the visual cue and the voiceprint cue on the unquantized LiMuSE (GC+CC). The results are listed in Table 8. We can see from the table that the performance decreases after discarding the voiceprint cue or the visual cue, especially the visual cue. LiMuSE (GC+CC-Vis) suffers from a larger performance gap compared to LiMuSE (GC+CC-Vp). There is about a 10dB performance gap when and a 7dB performance gap when between LiMuSE and LiMuSE (GC+CC-Vis). It suggests that the visual cue is more informative than the voiceprint cue. It can be further concluded that the voiceprint cue and the visual cue serve as complementary information for each other and introducing both of them into LiMuSE leads to the best performance compared with LiMuSE (GC+CC-Vp) and LiMuSE (GC+CC-Vis).
5 Conclusion
In this paper, we propose a lightweight multi-modal speaker extraction model incorporating the voiceprint and visual features as complementary cues. We propose to use the GC-equipped TCN as the basic module to compress the model size, which is used in the Context Codec module and the separation network in LiMuSE. Furthermore, we explore the impact of the quantization technologies. In an extreme case, the model size of LiMuSE is compressed to 0.19MB. Our experiments show that LiMuSE achieves promising results on MC_GRID dataset, and the quantized model achieves comparable performance with a compression rate as high as 180 compared with the baselines.
6 ACKNOWLEDGMENTS
This work was supported by the National Key Research and Development Program of China (2021ZD0201500).
References
- [1] Adelbert W Bronkhorst, “The cocktail party phenomenon: A review of research on speech intelligibility in multiple-talker conditions,” Acta Acustica united with Acustica, vol. 86, no. 1, pp. 117–128, 2000.
- [2] Dong Yu, Morten Kolbæk, Zheng-Hua Tan, and Jesper Jensen, “Permutation invariant training of deep models for speaker-independent multi-talker speech separation,” in ICASSP. IEEE, 2017, pp. 241–245.
- [3] Zhuo Chen, Yi Luo, and Nima Mesgarani, “Deep attractor network for single-microphone speaker separation,” ICASSP, pp. 246–250, 2017.
- [4] Chenglin Xu, Wei Rao, Eng Siong Chng, and Haizhou Li, “Spex: Multi-scale time domain speaker extraction network,” IEEE/ACM transactions on audio, speech, and language processing, vol. 28, pp. 1370–1384, 2020.
- [5] Quan Wang, Ignacio Lopez Moreno, Mert Saglam, Kevin Wilson, Alan Chiao, Renjie Liu, Yanzhang He, Wei Li, Jason Pelecanos, Marily Nika, and Alexander Gruenstein, “VoiceFilter-Lite: Streaming Targeted Voice Separation for On-Device Speech Recognition,” in Interspeech, 2020, pp. 2677–2681.
- [6] Meng Ge, Chenglin Xu, Longbiao Wang, Eng Siong Chng, Jianwu Dang, and Haizhou Li, “Multi-stage speaker extraction with utterance and frame-level reference signals,” in ICASSP. IEEE, 2021, pp. 6109–6113.
- [7] Yunzhe Hao, Jiaming Xu, Peng Zhang, and Bo Xu, “Wase: Learning when to attend for speaker extraction in cocktail party environments,” in ICASSP. IEEE, 2021, pp. 6104–6108.
- [8] Jian Wu, Yong Xu, Shi-Xiong Zhang, Lian-Wu Chen, Meng Yu, Lei Xie, and Dong Yu, “Time domain audio visual speech separation,” in ASRU. IEEE, 2019, pp. 667–673.
- [9] Rui Lu, Zhiyao Duan, and Changshui Zhang, “Audio–visual deep clustering for speech separation,” IEEE/ACM Transactions on Audio, Speech, and Language Processing, vol. 27, no. 11, pp. 1697–1712, 2019.
- [10] Ariel Ephrat, Inbar Mosseri, Oran Lang, Tali Dekel, Kevin Wilson, Avinatan Hassidim, William T Freeman, and Michael Rubinstein, “Looking to listen at the cocktail party: a speaker-independent audio-visual model for speech separation,” TOG, vol. 37, no. 4, pp. 1–11, 2018.
- [11] Peng Zhang, Jiaming Xu, Jing Shi, Yunzhe Hao, Lei Qin, and Bo Xu, “Audio-visual speech separation with visual features enhanced by adversarial training,” in IJCNN. IEEE, 2021.
- [12] Chenxing Li, Jiaming Xu, Nima Mesgarani, and Bo Xu, “Speaker and direction inferred dual-channel speech separation,” in ICASSP. IEEE, 2021, pp. 5779–5783.
- [13] Guanjun Li, Shan Liang, Shuai Nie, Wenju Liu, Meng Yu, Lianwu Chen, Shouye Peng, and Changliang Li, “Direction-aware speaker beam for multi-channel speaker extraction.,” in Interspeech, 2019, pp. 2713–2717.
- [14] Rongzhi Gu, Shi-Xiong Zhang, Yong Xu, Lianwu Chen, Yuexian Zou, and Dong Yu, “Multi-modal multi-channel target speech separation,” IEEE Journal of Selected Topics in Signal Processing, vol. 14, no. 3, pp. 530–541, 2020.
- [15] Andrew G Howard, Menglong Zhu, Bo Chen, Dmitry Kalenichenko, Weijun Wang, Tobias Weyand, Marco Andreetto, and Hartwig Adam, “Mobilenets: Efficient convolutional neural networks for mobile vision applications,” arXiv preprint arXiv:1704.04861, 2017.
- [16] Mark Sandler, Andrew Howard, Menglong Zhu, Andrey Zhmoginov, and Liang-Chieh Chen, “Mobilenetv2: Inverted residuals and linear bottlenecks,” in CVPR, 2018, pp. 4510–4520.
- [17] Yi Luo, Cong Han, and Nima Mesgarani, “Ultra-lightweight speech separation via group communication,” in ICASSP. IEEE, 2021, pp. 16–20.
- [18] Yi Luo, Cong Han, and Nima Mesgarani, “Group communication with context codec for lightweight source separation,” IEEE/ACM Transactions on Audio, Speech, and Language Processing, vol. 29, pp. 1752–1761, 2021.
- [19] Jian-Hao Luo, Jianxin Wu, and Weiyao Lin, “Thinet: A filter level pruning method for deep neural network compression,” in ICCV, 2017, pp. 5058–5066.
- [20] Song Han, Jeff Pool, John Tran, and William Dally, “Learning both weights and connections for efficient neural network,” Advances in neural information processing systems, vol. 28, 2015.
- [21] Geoffrey Hinton, Oriol Vinyals, and Jeffrey Dean, “Distilling the knowledge in a neural network,” in NIPS Deep Learning and Representation Learning Workshop, 2015.
- [22] Byeongho Heo, Minsik Lee, Sangdoo Yun, and Jin Young Choi, “Knowledge distillation with adversarial samples supporting decision boundary,” in AAAI, 2019, vol. 33, pp. 3771–3778.
- [23] Mohammad Rastegari, Vicente Ordonez, Joseph Redmon, and Ali Farhadi, “Xnor-net: Imagenet classification using binary convolutional neural networks,” in ECCV. Springer, 2016, pp. 525–542.
- [24] Shuang Wu, Guoqi Li, Feng Chen, and Luping Shi, “Training and inference with integers in deep neural networks,” in ICLR, 2018.
- [25] Shuchang Zhou, Yuxin Wu, Zekun Ni, Xinyu Zhou, He Wen, and Yuheng Zou, “Dorefa-net: Training low bitwidth convolutional neural networks with low bitwidth gradients,” arXiv preprint arXiv:1606.06160, 2016.
- [26] Yating Huang, Yunzhe Hao, Jiaming Xu, and Bo Xu, “Compressing speaker extraction model with ultra-low precision quantization and knowledge distillation,” Neural Networks, 2022.
- [27] Yi Luo and Nima Mesgarani, “Conv-tasnet: Surpassing ideal time–frequency magnitude masking for speech separation,” IEEE/ACM transactions on audio, speech, and language processing, vol. 27, no. 8, pp. 1256–1266, 2019.
- [28] Hervé Bredin, Ruiqing Yin, Juan Manuel Coria, Gregory Gelly, Pavel Korshunov, Marvin Lavechin, Diego Fustes, Hadrien Titeux, Wassim Bouaziz, and Marie-Philippe Gill, “pyannote.audio: neural building blocks for speaker diarization,” in ICASSP, Barcelona, Spain, May 2020.
- [29] Hang Zhou, Yu Liu, Ziwei Liu, Ping Luo, and Xiaogang Wang, “Talking face generation by adversarially disentangled audio-visual representation,” in AAAI, 2019.
- [30] Joon Son Chung, Andrew Senior, Oriol Vinyals, and Andrew Zisserman, “Lip reading sentences in the wild,” in CVPR, 2017, pp. 3444–3453.
- [31] Yandong Guo, Lei Zhang, Yuxiao Hu, Xiaodong He, and Jianfeng Gao, “Ms-celeb-1m: A dataset and benchmark for large-scale face recognition,” in ECCV, Bastian Leibe, Jiri Matas, Nicu Sebe, and Max Welling, Eds., Cham, 2016, pp. 87–102, Springer International Publishing.
- [32] Helen L Bear and Richard Harvey, “Phoneme-to-viseme mappings: the good, the bad, and the ugly,” Speech Communication, vol. 95, pp. 40–67, 2017.
- [33] Yong Yu, Xiaosheng Si, Changhua Hu, and Jianxun Zhang, “A review of recurrent neural networks: Lstm cells and network architectures,” Neural Computation, vol. 31, no. 7, pp. 1235–1270, 2019.
- [34] Yi Luo, Zhuo Chen, Nima Mesgarani, and Takuya Yoshioka, “End-to-end microphone permutation and number invariant multi-channel speech separation,” in ICASSP. IEEE, 2020, pp. 6394–6398.
- [35] Jie Hu, Li Shen, and Gang Sun, “Squeeze-and-excitation networks,” in CVPR, 2018, pp. 7132–7141.
- [36] Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun, “Delving deep into rectifiers: Surpassing human-level performance on imagenet classification,” in ICCV, 2015, pp. 1026–1034.
- [37] Doroteo T Toledano, María Pilar Fernández-Gallego, and Alicia Lozano-Diez, “Multi-resolution speech analysis for automatic speech recognition using deep neural networks: Experiments on timit,” PloS one, vol. 13, no. 10, pp. e0205355, 2018.
- [38] Jiwei Yang, Xu Shen, Jun Xing, Xinmei Tian, Houqiang Li, Bing Deng, Jianqiang Huang, and Xian-sheng Hua, “Quantization networks,” in CVPR, 2019, pp. 7308–7316.
- [39] Raghuraman Krishnamoorthi, “Quantizing deep convolutional networks for efficient inference: A whitepaper,” arXiv preprint arXiv:1806.08342, 2018.
- [40] Matthieu Courbariaux, Yoshua Bengio, and Jean-Pierre David, “Binaryconnect: Training deep neural networks with binary weights during propagations,” Advances in neural information processing systems, vol. 28, 2015.
- [41] Martin Cooke, Jon Barker, Stuart Cunningham, and Xu Shao, “An audio-visual corpus for speech perception and automatic speech recognition,” The Journal of the Acoustical Society of America, vol. 120, no. 5, pp. 2421–2424, 2006.
- [42] Lukas Drude, Jens Heitkaemper, Christoph Boeddeker, and Reinhold Haeb-Umbach, “SMS-WSJ: Database, performance measures, and baseline recipe for multi-channel source separation and recognition,” arXiv preprint arXiv:1910.13934, 2019.
- [43] Diederik P Kingma and Jimmy Ba, “Adam: A method for stochastic optimization,” in ICLR, 2015.
- [44] Zexu Pan, Ruijie Tao, Chenglin Xu, and Haizhou Li, “Muse: Multi-modal target speaker extraction with visual cues,” in ICASSP 2021-2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, 2021, pp. 6678–6682.
- [45] Joon Son Chung, Arsha Nagrani, and Andrew Zisserman, “Voxceleb2: Deep speaker recognition,” Proc. Interspeech 2018, pp. 1086–1090, 2018.
- [46] Triantafyllos Afouras, Joon Son Chung, Andrew Senior, Oriol Vinyals, and Andrew Zisserman, “Deep audio-visual speech recognition,” IEEE transactions on pattern analysis and machine intelligence, 2018.
- [47] Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun, “Deep residual learning for image recognition,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2016, pp. 770–778.