This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Source-Aware Neural Speech Coding for Noisy Speech Compression

Abstract

This paper introduces a novel neural network-based speech coding system that can process noisy speech effectively. The proposed source-aware neural audio coding (SANAC) system harmonizes a deep autoencoder-based source separation model and a neural coding system, so that it can explicitly perform source separation and coding in the latent space. An added benefit of this system is that the codec can allocate a different amount of bits to the underlying sources, so that the more important source sounds better in the decoded signal. We target a new use case where the user on the receiver side cares about the quality of the non-speech components in the speech communication, while the speech source still carries the most important information. Both objective and subjective evaluation tests show that SANAC can recover the original noisy speech better than the baseline neural audio coding system, which is with no source-aware coding mechanism, and two conventional codecs.

Index Terms—  Speech enhancement, speech coding, source separation

1 Introduction

Breakthroughs made in deep learning for the past decade have shown phenomenal performance improvements in various pattern recognition tasks, including media compression and coding. Seminal works are proposed in the lossy image compression domain, where autoencoders are a natural choice. With an autoencoder, the encoder part converts the input signal into a latent feature vector, followed by the decoder that recovers the original input [1, 2]. Compression is achieved when the number of bits used to represent the latent vector (or code) is smaller than that of the raw input signal. With increased computational complexity, the deep autoencoders have shown superior compression performance to traditional technology.

Neural speech coding is an emerging research area, too. Autoregressive models, such as WaveNet [3], have shown a transparent perceptual performance at a very low bitrate [4, 5], surpassing that of traditional coders. Another branch of neural speech coding systems takes a frame-by-frame approach, feeding time-domain waveform signals to an end-to-end autoencoding network. Kankanahalli proposes a simpler model that consists of fully convolutional layers to integrate dimension reduction, quantization, and entropy control tasks [6]. Cross-module residual learning (CMRL) inherits the convolutional pipeline and proposes a cascading structure, where multiple autoencoders are concatenated to work on the residual signal produced by the preceding ones [7]. In [8], CMRL is coupled with a trainable linear predictive coding (LPC) module as a pre-processor. It further improves the performance and lowered the model complexity down to 0.45 million parameters, eventually outperforming AMR-WB [9].

In this work, we widen the scope of speech coding applications by taking into account noisy speech as the input. Additional sound sources often accompany real-world speech. However, traditional speech codecs are mostly based on the speech production models [10, 9], thus lacking the ability to model the non-speech components mixed in the input signal. Efforts to address the problem have been partly reflected in the MPEG unified speech and audio coding (USAC) standard [11, 12]. USAC tackles speech signals in the mixture condition by switching between different tools defined for different kinds of signals, such as speech and music. However, the switching decision does not consider the mixed nature within the frame, which requires explicit source separation. Meanwhile, AMR-WB’s discontinuous transmission (DTX) mode also considers the mixed nature of input speech by deactivating the coding process for the non-speech periods [9]. Lombard et al. improved DTX by generating artificial comfort noise that smooths out the discontinuity [13]. However, for the frames where both speech and non-speech sources co-exist, it is difficult to effectively control the bitrate using DTX. Similar ideas have been used in transform coders for audio compression, where the dynamic bit allocation algorithm based on psychoacoustic models can create a spectral hole in low bitrate cases. Intelligent noise gap filling can alleviate the musical noise generated from this quantization process [14, 15], while it is to reduce the artifact generated from the coding algorithm, rather than to model the non-stationary noise source separately from the main source.

To that end, we propose source-aware neural audio coding (SANAC) to control the bit allocation to multiple sources differently. SANAC does not seek a speech-only reconstruction, e.g., by denoising the noisy input while coding it simultaneously [16]. Instead, we target the use case where the user still wants the code to convey the non-speech components to better understand the transmitter’s acoustic environment. We empirically show that the sources in the mixture can be assigned with unbalanced bitrates depending on their perceptual or applicational importance and entropy in the latent space, leading to a better objective and subjective quality.

2 Model Description

The proposed SANAC system harmonizes a source separation module into the neural coding system. Our model performs explicit source separation in the feature space to produce source-specific codes, which are subsequently quantized and decoded to recover the respective sources. The source-specific code vectors can be learned using a masking-based approaches as in TasNets [17, 18], while we propose to utilize the orthogonality assumption between the source-specific code vectors to drop the separator module in the TasNet architecture and reduce the encoder complexity. Soft-to-hard quantization [2] quantizes the real-valued source-specific codes. Bitrate control works on individual sources independently as well.

2.1 Orthogonal code vectors for separation

As a codec system, the model consists of an encoder that converts time-domain mixture frame 𝒙=N\bm{x}=\in\mathbb{R}^{N} into code vector 𝒛D\bm{z}\in\mathbb{R}^{D}: 𝒛enc(𝒙)\bm{z}\leftarrow{\mathcal{F}}_{\text{enc}}(\bm{x}). To marry the source separation concept, we assume KK mask vectors 𝒎(k)D\bm{m}^{(k)}\in\mathbb{R}^{D} that can decompose the code vector into KK components: k=1Kmd(k)=1\sum_{k=1}^{K}m^{(k)}_{d}=1. Note that md(k)m^{(k)}_{d} is the probability of dd-th code value belonging to kk-th source. In addition, we further assume that this probabilistic code assignments to the sources are actually determined by one-hot vectors, so that the masking process assigns each code value to only one source, i.e.,

md(k)={1 if argmaxjmd(j)=k0 otherwise.m_{d}^{(k)}=\left\{\begin{array}[]{rl}1&\text{ if }\operatornamewithlimits{\arg\max}_{j}m_{d}^{(j)}=k\\ 0&\text{ otherwise.}\end{array}\right. (1)

The TasNet models estimate similar masking vectors via a separate neural network module, which led to the state-of-the-art separation performance. In there, the estimated mask values are indeed somewhat drastically distributed to either near zero or one, making the masked code vectors nearly orthogonal from each other. However, the sigmoid-activated masks in the TasNet architecture and our probabilistic masks do not specifically assume a hard assignment.

From now on, we assume orthogonal code vectors per source as a result of hard masking, i.e., 𝒛(1)𝒛(2)𝒛(K)\bm{z}^{(1)}\!\perp\!\bm{z}^{(2)}\!\perp\!\cdots\!\perp\!\bm{z}^{(K)}, where 𝒛(k)=𝒎(k)𝒛\bm{z}^{(k)}=\bm{m}^{(k)}\odot\bm{z}, kk-th source’s code vector defined by the Hadamard product \odot between the mask and the mixture code.

The proposed orthogonality leads us to a meaningful structural innovation. Instead of estimating the mask vector for every input frame, we can use structured masking vectors that force the code values to be grouped into KK exclusive and consecutive subsets. For example, for a two-source case with D=8D=8, 𝒎(1)=[1,1,1,1,0,0,0,0]\bm{m}^{(1)}=[1,1,1,1,0,0,0,0]^{\top}. Hence, we can safely discard the masked-out elements (the latter four elements), by defining its truncated version as 𝐳(k)D/K\mathbf{z}^{(k)}\in\mathbb{R}^{D/K}. The concatenation of the truncated code vectors determines the final code vector: 𝒛=[𝐳(1),𝐳(2),,𝐳(K)]\bm{z}=[{\mathbf{z}^{(1)}}^{\top},{\mathbf{z}^{(2)}}^{\top},\cdots,{\mathbf{z}^{(K)}}^{\top}]^{\top}.

In practice, we implement the encoder as a 1-d convolutional neural network (CNN) whose output is a 2L×P2L\times P matrix, where 2L2L is the number of output channels (see Figure 1, where L=6L=6, P=256P=256). We collect the first LL channels of this feature map (dark red bars) as our codes for speech, i.e., 𝐳(1)\mathbf{z}^{(1)} corresponds to the vectorized version of the upper half of the feature map of size L×PL\times P, or LP=D/KLP=D/K, where DD should be an integer multiple of KK. The other half is for the noise source. Since the decoders are learned to predict individual sources, this implicit masking process can still work for source separation.

2.2 Soft-to-hard quantization

Quantization is a mapping process that replaces a continuous variable into its closest discrete representative. Since it is not a differentiable process, incorporating it into a neural network requires careful consideration. Soft-to-hard quantization showed successful performance both in image and speech compression models [2, 6, 7]. The idea is to formulate this cluster assignment process as a softmax classification during the feedforward process, which finds the nearest one among MM total representatives 𝝁m\bm{\mu}_{m} for the given code vector 𝒚\bm{y} as follows:

dm\displaystyle d_{m} =(𝒚||𝝁m),𝒑=Softmax(α𝒅),\displaystyle={\mathcal{E}}(\bm{y}||\bm{\mu}_{m}),\quad\bm{p}=\text{Softmax}(-\alpha\bm{d}), (2)
Testing: 𝒚¯\displaystyle\text{Testing: }\bar{\bm{y}} =𝝁argmaxmpm,Training: 𝒚¯=m=1Mpm𝝁m,\displaystyle=\bm{\mu}_{\operatornamewithlimits{\arg\max}_{m}p_{m}},\quad\text{Training: }\bar{\bm{y}}=\sum_{m=1}^{M}p_{m}\bm{\mu}_{m},

where the algorithm first computes the Euclidean distance vector 𝒅\bm{d} against all the representatives (i.e., the cluster means), whose negative value works like a similarity score for the softmax function. Using the softmax result, the probability vector 𝒑\bm{p} of the cluster membership, we can construct the quantized code vector 𝒚¯\bar{\bm{y}}: during test time, simply choosing the closest one will do a proper quantization. However, since the argmax\operatornamewithlimits{\arg\max} operation is not differentiable, for training, we do a convex combination of the cluster centroids to represent the quantized code, as a differentiable surrogate of the hard assignment process. The discrepancy between training and testing is reduced by controlling the scaling hyperparameter α\alpha, which makes the softmax probabilities more drastic once it is large enough (e.g., a one-hot vector in the extreme case). Note that it also learns the cluster centroids 𝝁\bm{\mu} as a part of the learnable network parameters rather than employing a separate clustering process to define them.

In previous work the quantization has been on scalar variables, i.e., 𝒚1\bm{y}\in\mathbb{R}^{1} [6, 7, 8]. In this work, the soft-to-hard quantization performs vector quantization (VQ). We denote the CNN encoder output by 𝒁2L×P\bm{Z}\in\mathbb{R}^{2L\times P}, which consists of K=2K=2 code blocks: 𝒁=[𝒁(1);𝒁(2)]\bm{Z}=[\bm{Z}^{(1)};\bm{Z}^{(2)}]. Then, each code vector for quantization is defined by the pp-th feature spanning over LL channels: 𝒚=𝒁1:L,p(k)\bm{y}=\bm{Z}_{1:L,p}^{(k)}, having L=6L=6 as the VQ dimension in our case.

2.3 Source-wise entropy control

The theoretical lower bound of the bitrate, as a result of Huffman coding, can be defined by the entropy of the quantized codes. The frequency of the cluster means defines the entropy of the source-specific codes: (𝝁(k))=m=1Mqm(k)logqm(k){\mathcal{H}}\big{(}\bm{\mu}^{(k)}\big{)}=-\sum_{m=1}^{M}q_{m}^{(k)}\log q_{m}^{(k)}, where qm(k)q_{m}^{(k)} denotes the frequency of mm-th mean for the kk-th source. Meanwhile, the entropy for the code of the mixture signal is smaller than or equal to the sum of the entropy of all sources: (𝝁¨)k=1K(𝝁(k)){\mathcal{H}}(\ddot{\bm{\mu}})\leq\sum_{k=1}^{K}{\mathcal{H}}\big{(}\bm{\mu}^{(k)}\big{)}, where 𝝁¨\ddot{\bm{\mu}} is the set of quantization vector centroids learned in a source-agnostic way, i.e., directly from the mixture signals. Therefore, in theory, SANAC cannot achieve a better coding gain than a codec that works directly on the mixture.

However, SANAC can still benefit from the source-wise coding, especially by exploiting the perceptual factors. Our main assumption in this work is that the perceptual importance differs from source-by-source, leading to a coding system that can assign different bitrates to different sources. For noisy speech, for example, we will try to assign more bits to the speech source. Consequently, although the user eventually listens to the recovered mixture of speech and noise (a) the perceptual quality of the speech component is relatively higher (b) the codec can achieve a better coding gain if the noise source’ statistical characteristics is robust to low bitrates.

Our argument is based on the codec’s ability to control the entropy of the source-specific codes. In SANAC, we adopt the entropy control mechanism proposed in [2], but by setting up a per-source loss between the target ξ(k)\xi^{(k)} and the actual entropy values: (ξ(k)(𝝁(k)))2\big{(}\xi^{(k)}-{\mathcal{H}}\big{(}\bm{\mu}^{(k)}\big{)}\big{)}^{2}. While this loss does not guarantee the exact bitrate during the test time, in practice, we observe that the actual bitrate is not significantly different from the target.

Refer to caption
Fig. 1: Schematic diagram of source-aware speech coding. At Stage 1, the Stage 2 and 3 modules are not trained, while Stage 2 updates all parameters except for the Stage 3 modules. At Stage 3, all parameters are optimized.
Refer to caption
Fig. 2: Schematic diagram of DNN-based baseline model

2.4 Decoding and the final loss

The source-specific truncated codes, after the quantization, 𝒁¯(k)\bar{\bm{Z}}^{(k)}, are fed to the decoder part of the network. The decoder function works similarly to Conv-TasNet [18] in that the decoder runs KK times to predict KK individual source reconstructions from KK sourse-specific feature maps as the input. However, SANAC’s decoding is different from Conv-TasNet’s as the decoder input is the quantized codes. In addition, our model cares about the quality of the recovered mixture, not only the separation quality.

Our training loss considers all these goals, consisting of the main mean squared error (MSE)-based reconstruction term and the entropy control terms. More specifically, for the noisy speech case 𝒙=𝒔+𝒏\bm{x}=\bm{s}+\bm{n} (k=1k=1 for speech and k=2k=2 for noise), the MSE loss is for the speech source reconstruction 𝒔^\hat{\bm{s}} and the mixture reconstruction 𝒙^\hat{\bm{x}}, while the noise source reconstruction 𝒏^\hat{\bm{n}} is implied in there. We regularize the total entropy as well as the ratio between the two source-wise entropy values:

=λMSE(MSE(𝒔||𝒔^)+MSE(𝒙||𝒙^))+λEntTot(ξ(𝝁(1))(𝝁(2)))2+λRatio(ψ(𝝁(1))(𝝁(2)))2,\begin{split}{\mathcal{L}}&=\lambda_{\text{MSE}}\big{(}{\mathcal{E}}_{\text{MSE}}(\bm{s}||\hat{\bm{s}})+{\mathcal{E}}_{\text{MSE}}(\bm{x}||\hat{\bm{x}})\big{)}\\ &+\lambda_{\text{EntTot}}\Big{(}\xi-{\mathcal{H}}\big{(}\bm{\mu}^{(1)}\big{)}-{\mathcal{H}}\big{(}\bm{\mu}^{(2)}\big{)}\Big{)}^{2}\\ &+\lambda_{\text{Ratio}}\left(\psi-\frac{{\mathcal{H}}\big{(}\bm{\mu}^{(1)}\big{)}}{{\mathcal{H}}\big{(}\bm{\mu}^{(2)}\big{)}}\right)^{2},\end{split} (3)

where ξ\xi and ψ\psi are the target total entropy and the target ratio, respectively.

3 Experiment

3.1 Dataset

500 and 50 utterances are randomly selected from training and test set of TIMIT corpus [19]. We generate 10 contaminated samples out of each clean utterance by adding 10 different non-stationary background sources, {bird singing, casino, cicadas, typing, chip eating, frogs, jungle, machine gun, motorcycle, ocean} used in [20]. Every contaminated speech waveform was segmented into frames of 512 samples (32ms), with overlap of 64 samples. We apply a Hann window of size 128 samples only to the overlapping regions. Since there are 16000/44816000/448 frames per second and each frame produces PP code vectors for VQ, for the entropy of a source-specific codebook ξ\xi, the bitrate is 16000Pξ/48816000P\xi/488, e.g., 9.14kbps when P=256P=256 and ξ=1\xi=1.

3.2 Training process

Adam optimizer with an initial learning rate 0.0001 trains the models [21]. Both SANAC and the baseline are trained in three stages. Every jump to next stage is triggered when the validation loss stops improving in 3 consecutive epochs. We stop the training updates after validation loss does not improve for 20 epochs.

  • Stage 1: For the first three epochs, the model trains the encoder to separate the input into the speech and background sources, that are represented by the two orthogonal code vectors. No quantization is involved in yet, but this stage better initializes the parameters for the quantization process. The encoder consists of a few bottleneck blocks that are commonly used in ResNet [22]. With the bottleneck structure, the input and output feature maps can be connected via an identity shortcut with less filters to learn in between. The encoder module also employs a 1-d convolution layer to downsample the feature map from 512 to 256, followed by another bottleneck block and a channel changing layer to yield two sets of code maps of 6×2566\times 256 each. The learned source-specific feature maps are fed directly to the channel changer with no quantization, followed by an upsampler and the final decoder. In terms of upsampling, we interlace two adjacent channels into one, doubling the number of features up to 512 while halving the number of channels as introduced in [23], and then adopted for neural speech coding [6, 7].

  • Stage 2: In this stage, the model starts to quantize the encoder output using the soft-to-hard VQ mechanism. The VQ is done with M=128M=128 centroids. We set the scale of the softmax function α=10\alpha=10, and increase it exponentially until it reaches 500 to gradually introduce more hardness to the softmax function. Meanwhile, the other modules in the network are also updated accordingly to absorb the quantization error. As the model stabilizes, we introduce entropy control terms into the loss function by setting up the regularization weights λEntTot=1/5\lambda_{\text{EntTot}}=1/5 and λRatio=1/60\lambda_{\text{Ratio}}=1/60. We set ξ\xi to be 1, 2, and 3, which correspond to three bitrates 9.14, 18.29, and 27.43kbps. The target ratio between speech and noise bitrates is trainable, which we set to be ψ=3\psi=3 at the beginning.

  • Stage 3: Finally, the feature maps go through a few more ResNet blocks that strengthen the separation (the “Separator” module in Figure 1), followed by the decoder that runs twice for both sources. The “Separator” and “Decoder” consist of two bottleneck blocks with different input channels, 60 and 30, respectively. “Decoder” needs an additional channel changer.

  • The Baseline: Our baseline system shown in Fig. 2 is similar to the proposed architecture, except that it discards the first separation before coding, meaning that the codec system is not aware of the different sources. Hence, the baseline only generates one type of codes representing the mixture signal without any control of the entropy ratio in the loss function. We still use the two-headed decoder architecture to benefit from the two reconstruction loss terms.

Table 1: MOS-LQO results
0dB                            Comparing systems
Bitrate Baseline Proposed AMR-WB Opus
9\sim 9kbps
1.53,1.031.53,1.03 1.66,1.021.66,1.02 1.50,1.161.50,1.16 1.66,1.101.66,1.10
18\sim 18kbps
1.54,1.031.54,1.03 1.56,1.021.56,1.02 1.54,1.151.54,1.15 1.59,1.131.59,1.13
27\sim 27kbps
1.37,1.031.37,1.03 1.37,1.031.37,1.03 1.49,1.141.49,1.14 1.52,1.121.52,1.12
5dB
9\sim 9kbps
1.52,1.021.52,1.02 1.53,1.021.53,1.02 1.31,1.381.31,1.38 1.39,1.101.39,1.10
18\sim 18kbps
1.38,1.021.38,1.02 1.47,1.021.47,1.02 1.67,1.351.67,1.35 1.40,1.321.40,1.32
27\sim 27kbps
1.74,1.021.74,1.02 1.70,1.061.70,1.06 1.48,1.331.48,1.33 1.21,1.281.21,1.28
Refer to caption
(a) STOI of recovered mixtures
Refer to caption
(b) SiSDR of recovered mixtures
Fig. 3: Objective evaluation of the neural network-based models
Table 2: Subjective results from the A/B test
Bitrates 9\sim 9kbps 18\sim 18kbps 27\sim 27kbps
Preference rate 96.3%96.3\% 80.0%80.0\% 56.3%56.3\%
Standard deviation 0.700.70 1.01.0 1.221.22

3.3 Results and analysis

We first show the average MOS-LQO (PESQ) score for SANAC, baseline and two standard codecs (AMR-WB [9] and Opus [24]), at 3 different target bitrates in Table 1. We evaluate the reconstruction of the input noisy speech. For each decoded sample, both the uncompressed mixture input and its corresponding clean speech source are considered as the reference, yielding the left and right hand scores, respectively. Hence, the right hand score measures the speech enhancement performance, while the left hand score indicates the level of overall reconstruction quality. Since our neural codecs are not designed to degrade noise source during compression, the right hand scores are lower than those from standard codecs where the added noise is somewhat suppressed. In terms of the left hand score, neural codecs outperform those counterparts in most cases. It is known, however, that PESQ is not well defined for the perceptual evaluation between noisy speech signals, which to a certain degree explains the very similar left hand scores from the baseline and SANAC.

Additionally, the baseline and proposed SANAC are evaluated based on the scale-invariant signal-to-distortion ratio (SiSDR) [25] and short-time objective intelligibility (STOI) [26]. The clean speech is used as the reference when computing STOI. Similar comparison results are observed for both STOI (Fig.3(a)) and SiSDR (Fig.3(b)): overall, SANAC achieves higher scores in most cases. For STOI, we find that the performance gap becomes more prominent in the lower bitrates, thanks to the more speech-oriented setup we used for SANAC. With respect to SiSDR, SANAC shows better mixture reconstruction performance, too. The margin against the baseline is more noticeable when the SNR level of the input mixture is higher.

Finally, we conduct a subjective listening test with eight audio experts participated. The A/B tests are designed to compare our own baseline and SANAC; as a controlled experiment, where all other system aspects are fixed, the test focuses on the impact of the source-aware coding scheme. The test consists of three sessions with different bitrates (9\sim 9, 18\sim 18, and 27\sim 27kbps). With the uncompressed mixture signal as the reference, listeners are asked to designate the sample from 2 competing systems that sounds more similar to the reference. Results are presented in Table 2, where the preference rate denotes how likely an average listener prefers SANAC. We also report the standard deviation to measure the variation of the preference rate. In all cases, SANAC is favored by most listeners. As the bitrate gets lower, the listener tends to have a higher and more determined preference rate for SANAC over the baseline.

4 Conclusion

In this work, we proposed SANAC for source-aware neural speech coding. In the model, we harmonized a Conv-TasNet-like masking-based separation approach with an end-to-end neural speech coding network, so that the system can produce source-specific codes. SANAC showcased superior performance in both objective and subjective tests to the baseline source-agnostic model with a similar architecture. We believe that SANAC opens a new possibility of widening audio coding on mixture signals by being able to control individual sources differently. The sound examples and source code are available online111https://saige.sice.indiana.edu/research-projects/sanac.

References

  • [1] L. Theis, W. Shi, A. Cunningham, and F. Huszar, “Lossy image compression with compressive autoencoders,” 2017.
  • [2] E. Agustsson, F. Mentzer, M. Tschannen, L. Cavigelli, R. Timofte, L. Benini, and L. V. Gool, “Soft-to-hard vector quantization for end-to-end learning compressible representations,” in Advances in Neural Information Processing Systems (NIPS), 2017, pp. 1141–1151.
  • [3] A. van den Oord, S. Dieleman, H. Zen, K. Simonyan, O. Vinyals, A. Graves, N. Kalchbrenner, A. Senior, and K. Kavukcuoglu, “Wavenet: A generative model for raw audio,” arXiv preprint arXiv:1609.03499, 2016.
  • [4] W. B. Kleijn, F. S. C. Lim, A. Luebs, J. Skoglund, F. Stimberg, Q. Wang, and T. C. Walters, “WaveNet based low rate speech coding,” in Proceedings of the IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP), 2018, pp. 676–680.
  • [5] J.-M. Valin and J. Skoglund, “A real-time wideband neural vocoder at 1.6 kb/s using LPCNet,” in Proceedings of the Annual Conference of the International Speech Communication Association (Interspeech), 2019.
  • [6] S. Kankanahalli, “End-to-end optimized speech coding with deep neural networks,” in Proceedings of the IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP), 2018.
  • [7] K. Zhen, J. Sung, M. S. Lee, S. Beack, and M. Kim, “Cascaded cross-module residual learning towards lightweight end-to-end speech coding,” in Proceedings of the Annual Conference of the International Speech Communication Association (Interspeech), 2019.
  • [8] K. Zhen, M. S. Lee, J. Sung, S. Beack, and M Kim, “Efficient and scalable neural residual waveform coding with collaborative quantization,” in Proceedings of the IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP), 2020.
  • [9] B. Bessette, R. Salami, R. Lefebvre, M. Jelinek, J. Rotola-Pukkila, J. Vainio, H. Mikkola, and K. Jarvinen, “The adaptive multirate wideband speech codec (AMR-WB),” IEEE Transactions on Speech and Audio Processing, vol. 10, no. 8, pp. 620–636, 2002.
  • [10] M. Schroeder and B. Atal, “Code-excited linear prediction (CELP): High-quality speech at very low bit rates,” in Proceedings of the IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP), 1985, vol. 10, pp. 937–940.
  • [11] ISO/IEC DIS 23003-3, “Information technology – mpeg audio technologies – part 3: Unified speech and audio coding,” 2011.
  • [12] ISO/IEC 14496-3:2009/PDAM 3, “Transport of unified speech and audio coding (USAC),” 2011.
  • [13] A. Lombard, S. Wilde, E. Ravelli, S. Döhla, G. Fuchs, and M. Dietz, “Frequency-domain comfort noise generation for discontinuous transmission in evs,” in Proceedings of the IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP), 2015, pp. 5893–5897.
  • [14] J. Herre, J. Hilpert, A. Kuntz, and J. Plogsties, “Mpeg-h audio–the new standard for universal spatial/3D audio coding,” Journal of Audio Engineering Society, vol. 62, no. 12, pp. 821–830, 2015.
  • [15] S. Disch, A. Niedermeier, C. R. Helmrich, C. Neukam, K. Schmidt, R. Geiger, J. Lecomte, F¿ Ghido, F. Nagel, and B. Edler, “Intelligent gap filling in perceptual transform coding of audio,” in Audio Engineering Society Convention 141, Sep. 2016.
  • [16] F. S. C. Lim, W. Bastiaan Kleijn, M. Chinen, and J. Skoglund, “Robust low rate speech coding based on cloned networks and wavenet,” in Proceedings of the IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP), 2020, pp. 6769–6773.
  • [17] Y. Luo and N. Mesgarani, “Tasnet: time-domain audio separation network for real-time, single-channel speech separation,” in Proceedings of the IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP), 2018.
  • [18] Y. Luo and N. Mesgarani, “Conv-TasNet: Surpassing ideal time–frequency magnitude masking for speech separation,” IEEE/ACM Transactions on Audio, Speech, and Language Processing, vol. 27, no. 8, pp. 1256–1266, 2019.
  • [19] J. S. Garofolo, L. F. Lamel, W. M. Fisher, J. G. Fiscus, D. S. Pallett, N. L. Dahlgren, and V. Zue, “TIMIT acoustic-phonetic continuous speech corpus,” Linguistic Data Consortium, Philadelphia, 1993.
  • [20] Z. Duan, G. J. Mysore, and P. Smaragdis, “Online PLCA for real-time semi-supervised source separation,” in Proceedings of the International Conference on Latent Variable Analysis and Signal Separation (LVA/ICA), 2012, pp. 34–41.
  • [21] Diederik P. Kingma and Jimmy Ba, “Adam: A method for stochastic optimization,” 2014.
  • [22] K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning for image recognition,” in Proceedings of the IEEE International Conference on Computer Vision and Pattern Recognition (CVPR), 2016, pp. 770–778.
  • [23] W. Shi, J. Caballero, F. Huszár, J. Totz, A. P. Aitken, R. Bishop, D. Rueckert, and Z. Wang, “Real-time single image and video super-resolution using an efficient sub-pixel convolutional neural network,” in Proceedings of the IEEE International Conference on Computer Vision and Pattern Recognition (CVPR), 2016, pp. 1874–1883.
  • [24] J. M. Valin, K. Vos, and T. Terriberry, “Definition of the opus audio codec,” IETF, September, 2012.
  • [25] J. Le Roux, S. Wisdom, H. Erdogan, and J. R. Hershey, “SDR - half-baked or well done?,” arXiv preprint arXiv:1811.02508, 2018.
  • [26] C. H. Taal, R. C. Hendriks, R. Heusdens, and J. Jensen, “A short-time objective intelligibility measure for time-frequency weighted noisy speech,” in Proceedings of the IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP), 2010.