This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

\stackMath

Cascaded Cross-Module Residual Learning towards Lightweight End-to-End Speech Coding

Kai Zhen,   Jongmo Sung,  Mi Suk Lee,  Seungkwon Beack,
Minje Kim
This work was supported by the Institute for Information and Communications Technology Promotion (IITP) funded by the Korea government (MSIT) under Grant 2017-0-00072 (Development of Audio/Video Coding and Light Field Media Fundamental Technologies for Ultra Realistic Tera-Media). Kai Zhen is with the Department of Computer Science and Cognitive Science Program at Indiana University, Bloomington, IN 47408 USA. Jongmo Sung, Mi Suk Lee, and Seungkwon Beack are with Electronics and Telecommunications Research Institute, Daejeon, Korea 34129. Minje Kim is with the Dept. of Intelligent Systems Engineering at Indiana University (e- mails: [email protected], [email protected], [email protected], [email protected], [email protected]).Manuscript received March XX, 2021; revised YYY ZZZ, 2021.

Cross Module Residual Learning And Collaborative Quantization for Efficient Neural Waveform Coding

Kai Zhen,   Jongmo Sung,  Mi Suk Lee,  Seungkwon Beack,
Minje Kim
This work was supported by the Institute for Information and Communications Technology Promotion (IITP) funded by the Korea government (MSIT) under Grant 2017-0-00072 (Development of Audio/Video Coding and Light Field Media Fundamental Technologies for Ultra Realistic Tera-Media). Kai Zhen is with the Department of Computer Science and Cognitive Science Program at Indiana University, Bloomington, IN 47408 USA. Jongmo Sung, Mi Suk Lee, and Seungkwon Beack are with Electronics and Telecommunications Research Institute, Daejeon, Korea 34129. Minje Kim is with the Dept. of Intelligent Systems Engineering at Indiana University (e- mails: [email protected], [email protected], [email protected], [email protected], [email protected]).Manuscript received March XX, 2021; revised YYY ZZZ, 2021.

Scalable and Efficient Neural Waveform Coding with Cascaded Coding and Collaborative Quantization

Kai Zhen,   Jongmo Sung,  Mi Suk Lee,  Seungkwon Beack,
Minje Kim
This work was supported by the Institute for Information and Communications Technology Promotion (IITP) funded by the Korea government (MSIT) under Grant 2017-0-00072 (Development of Audio/Video Coding and Light Field Media Fundamental Technologies for Ultra Realistic Tera-Media). Kai Zhen is with the Department of Computer Science and Cognitive Science Program at Indiana University, Bloomington, IN 47408 USA. Jongmo Sung, Mi Suk Lee, and Seungkwon Beack are with Electronics and Telecommunications Research Institute, Daejeon, Korea 34129. Minje Kim is with the Dept. of Intelligent Systems Engineering at Indiana University (e- mails: [email protected], [email protected], [email protected], [email protected], [email protected]).Manuscript received March XX, 2021; revised YYY ZZZ, 2021.

Neural Waveform Coding with Cascaded Coding And Collaborative Quantization

Kai Zhen,   Jongmo Sung,  Mi Suk Lee,  Seungkwon Beack,
Minje Kim
This work was supported by the Institute for Information and Communications Technology Promotion (IITP) funded by the Korea government (MSIT) under Grant 2017-0-00072 (Development of Audio/Video Coding and Light Field Media Fundamental Technologies for Ultra Realistic Tera-Media). Kai Zhen is with the Department of Computer Science and Cognitive Science Program at Indiana University, Bloomington, IN 47408 USA. Jongmo Sung, Mi Suk Lee, and Seungkwon Beack are with Electronics and Telecommunications Research Institute, Daejeon, Korea 34129. Minje Kim is with the Dept. of Intelligent Systems Engineering at Indiana University (e- mails: [email protected], [email protected], [email protected], [email protected], [email protected]).Manuscript received March XX, 2021; revised YYY ZZZ, 2021.

Revitalizing Multistage Quantization And Linear Prediction For Efficient Neural Waveform Coding

Kai Zhen,   Jongmo Sung,  Mi Suk Lee,  Seungkwon Beack,
Minje Kim
This work was supported by the Institute for Information and Communications Technology Promotion (IITP) funded by the Korea government (MSIT) under Grant 2017-0-00072 (Development of Audio/Video Coding and Light Field Media Fundamental Technologies for Ultra Realistic Tera-Media). Kai Zhen is with the Department of Computer Science and Cognitive Science Program at Indiana University, Bloomington, IN 47408 USA. Jongmo Sung, Mi Suk Lee, and Seungkwon Beack are with Electronics and Telecommunications Research Institute, Daejeon, Korea 34129. Minje Kim is with the Dept. of Intelligent Systems Engineering at Indiana University (e- mails: [email protected], [email protected], [email protected], [email protected], [email protected]).Manuscript received March XX, 2021; revised YYY ZZZ, 2021.

Cascaded Coding With Collaborative Quantization for Efficient Neural Waveform Compression

Kai Zhen,   Jongmo Sung,  Mi Suk Lee,  Seungkwon Beack,
Minje Kim
This work was supported by the Institute for Information and Communications Technology Promotion (IITP) funded by the Korea government (MSIT) under Grant 2017-0-00072 (Development of Audio/Video Coding and Light Field Media Fundamental Technologies for Ultra Realistic Tera-Media). Kai Zhen is with the Department of Computer Science and Cognitive Science Program at Indiana University, Bloomington, IN 47408 USA. Jongmo Sung, Mi Suk Lee, and Seungkwon Beack are with Electronics and Telecommunications Research Institute, Daejeon, Korea 34129. Minje Kim is with the Dept. of Intelligent Systems Engineering at Indiana University (e- mails: [email protected], [email protected], [email protected], [email protected], [email protected]).Manuscript received March XX, 2021; revised YYY ZZZ, 2021.

Efficient Speech Waveform Coding With Residual Cascading And Collaboratively Quantized LPC Preprocessing

Kai Zhen,   Jongmo Sung,  Mi Suk Lee,  Seungkwon Beack,
Minje Kim
This work was supported by the Institute for Information and Communications Technology Promotion (IITP) funded by the Korea government (MSIT) under Grant 2017-0-00072 (Development of Audio/Video Coding and Light Field Media Fundamental Technologies for Ultra Realistic Tera-Media). Kai Zhen is with the Department of Computer Science and Cognitive Science Program at Indiana University, Bloomington, IN 47408 USA. Jongmo Sung, Mi Suk Lee, and Seungkwon Beack are with Electronics and Telecommunications Research Institute, Daejeon, Korea 34129. Minje Kim is with the Dept. of Intelligent Systems Engineering at Indiana University (e- mails: [email protected], [email protected], [email protected], [email protected], [email protected]).Manuscript received March XX, 2021; revised YYY ZZZ, 2021.

Collaboratively Quantized LPC With Residual Cascading for Efficient Speech Waveform Coding

Kai Zhen,   Jongmo Sung,  Mi Suk Lee,  Seungkwon Beack,
Minje Kim
This work was supported by the Institute for Information and Communications Technology Promotion (IITP) funded by the Korea government (MSIT) under Grant 2017-0-00072 (Development of Audio/Video Coding and Light Field Media Fundamental Technologies for Ultra Realistic Tera-Media). Kai Zhen is with the Department of Computer Science and Cognitive Science Program at Indiana University, Bloomington, IN 47408 USA. Jongmo Sung, Mi Suk Lee, and Seungkwon Beack are with Electronics and Telecommunications Research Institute, Daejeon, Korea 34129. Minje Kim is with the Dept. of Intelligent Systems Engineering at Indiana University (e- mails: [email protected], [email protected], [email protected], [email protected], [email protected]).Manuscript received March XX, 2021; revised YYY ZZZ, 2021.

Deep Learning in Speech Waveform Coding

Kai Zhen,   Jongmo Sung,  Mi Suk Lee,  Seungkwon Beack,
Minje Kim
This work was supported by the Institute for Information and Communications Technology Promotion (IITP) funded by the Korea government (MSIT) under Grant 2017-0-00072 (Development of Audio/Video Coding and Light Field Media Fundamental Technologies for Ultra Realistic Tera-Media). Kai Zhen is with the Department of Computer Science and Cognitive Science Program at Indiana University, Bloomington, IN 47408 USA. Jongmo Sung, Mi Suk Lee, and Seungkwon Beack are with Electronics and Telecommunications Research Institute, Daejeon, Korea 34129. Minje Kim is with the Dept. of Intelligent Systems Engineering at Indiana University (e- mails: [email protected], [email protected], [email protected], [email protected], [email protected]).Manuscript received March XX, 2021; revised YYY ZZZ, 2021.

Residual Cascading With Collaborative Quantization Towards Efficient Blockwise Neural Speech Coding

Kai Zhen,   Jongmo Sung,  Mi Suk Lee,  Seungkwon Beack,
Minje Kim
This work was supported by the Institute for Information and Communications Technology Promotion (IITP) funded by the Korea government (MSIT) under Grant 2017-0-00072 (Development of Audio/Video Coding and Light Field Media Fundamental Technologies for Ultra Realistic Tera-Media). Kai Zhen is with the Department of Computer Science and Cognitive Science Program at Indiana University, Bloomington, IN 47408 USA. Jongmo Sung, Mi Suk Lee, and Seungkwon Beack are with Electronics and Telecommunications Research Institute, Daejeon, Korea 34129. Minje Kim is with the Dept. of Intelligent Systems Engineering at Indiana University (e- mails: [email protected], [email protected], [email protected], [email protected], [email protected]).Manuscript received March XX, 2021; revised YYY ZZZ, 2021.

Scalable Neural Waveform Coding for Speech Compression: A Lightweight Design

Kai Zhen,   Jongmo Sung,  Mi Suk Lee,  Seungkwon Beack,
Minje Kim
This work was supported by the Institute for Information and Communications Technology Promotion (IITP) funded by the Korea government (MSIT) under Grant 2017-0-00072 (Development of Audio/Video Coding and Light Field Media Fundamental Technologies for Ultra Realistic Tera-Media). Kai Zhen is with the Department of Computer Science and Cognitive Science Program at Indiana University, Bloomington, IN 47408 USA. Jongmo Sung, Mi Suk Lee, and Seungkwon Beack are with Electronics and Telecommunications Research Institute, Daejeon, Korea 34129. Minje Kim is with the Dept. of Intelligent Systems Engineering at Indiana University (e- mails: [email protected], [email protected], [email protected], [email protected], [email protected]).Manuscript received March XX, 2021; revised YYY ZZZ, 2021.

Neural Waveform Coding: A Lightweight Design

Kai Zhen,   Jongmo Sung,  Mi Suk Lee,  Seungkwon Beack,
Minje Kim
This work was supported by the Institute for Information and Communications Technology Promotion (IITP) funded by the Korea government (MSIT) under Grant 2017-0-00072 (Development of Audio/Video Coding and Light Field Media Fundamental Technologies for Ultra Realistic Tera-Media). Kai Zhen is with the Department of Computer Science and Cognitive Science Program at Indiana University, Bloomington, IN 47408 USA. Jongmo Sung, Mi Suk Lee, and Seungkwon Beack are with Electronics and Telecommunications Research Institute, Daejeon, Korea 34129. Minje Kim is with the Dept. of Intelligent Systems Engineering at Indiana University (e- mails: [email protected], [email protected], [email protected], [email protected], [email protected]).Manuscript received March XX, 2021; revised YYY ZZZ, 2021.

DSP Adapted Neural Waveform Coding

Kai Zhen,   Jongmo Sung,  Mi Suk Lee,  Seungkwon Beack,
Minje Kim
This work was supported by the Institute for Information and Communications Technology Promotion (IITP) funded by the Korea government (MSIT) under Grant 2017-0-00072 (Development of Audio/Video Coding and Light Field Media Fundamental Technologies for Ultra Realistic Tera-Media). Kai Zhen is with the Department of Computer Science and Cognitive Science Program at Indiana University, Bloomington, IN 47408 USA. Jongmo Sung, Mi Suk Lee, and Seungkwon Beack are with Electronics and Telecommunications Research Institute, Daejeon, Korea 34129. Minje Kim is with the Dept. of Intelligent Systems Engineering at Indiana University (e- mails: [email protected], [email protected], [email protected], [email protected], [email protected]).Manuscript received March XX, 2021; revised YYY ZZZ, 2021.

Scalable and Efficient Neural Speech Coding

Kai Zhen,   Jongmo Sung,  Mi Suk Lee,  Seungkwon Beack,
Minje Kim
This work was supported by the Institute for Information and Communications Technology Promotion (IITP) funded by the Korea government (MSIT) under Grant 2017-0-00072 (Development of Audio/Video Coding and Light Field Media Fundamental Technologies for Ultra Realistic Tera-Media). Kai Zhen is with the Department of Computer Science and Cognitive Science Program at Indiana University, Bloomington, IN 47408 USA. Jongmo Sung, Mi Suk Lee, and Seungkwon Beack are with Electronics and Telecommunications Research Institute, Daejeon, Korea 34129. Minje Kim is with the Dept. of Intelligent Systems Engineering at Indiana University (e- mails: [email protected], [email protected], [email protected], [email protected], [email protected]).Manuscript received March XX, 2021; revised YYY ZZZ, 2021.

Scalable and Efficient Neural Speech Coding:
A Hybrid Design

Kai Zhen,   Jongmo Sung,  Mi Suk Lee,  Seungkwon Beack,
Minje Kim
This work was supported by the Institute for Information and Communications Technology Promotion (IITP) funded by the Korea government (MSIT) under Grant 2017-0-00072 (Development of Audio/Video Coding and Light Field Media Fundamental Technologies for Ultra Realistic Tera-Media). Kai Zhen is with the Department of Computer Science and Cognitive Science Program at Indiana University, Bloomington, IN 47408 USA. Jongmo Sung, Mi Suk Lee, and Seungkwon Beack are with Electronics and Telecommunications Research Institute, Daejeon, Korea 34129. Minje Kim is with the Dept. of Intelligent Systems Engineering at Indiana University (e- mails: [email protected], [email protected], [email protected], [email protected], [email protected]).Manuscript received March XX, 2021; revised YYY ZZZ, 2021.
Abstract

We present a scalable and efficient neural waveform coding system for speech compression. We formulate the speech coding problem as an autoencoding task, where a convolutional neural network (CNN) performs encoding and decoding as a neural waveform codec (NWC) during its feedforward routine. The proposed NWC also defines quantization and entropy coding as a trainable module, so the coding artifacts and bitrate control are handled during the optimization process. We achieve efficiency by introducing compact model components to NWC, such as gated residual networks and depthwise separable convolution. Furthermore, the proposed models are with a scalable architecture, cross-module residual learning (CMRL), to cover a wide range of bitrates. To this end, we employ the residual coding concept to concatenate multiple NWC autoencoding modules, where each NWC module performs residual coding to restore any reconstruction loss that its preceding modules have created. CMRL can scale down to cover lower bitrates as well, for which it employs linear predictive coding (LPC) module as its first autoencoder. The hybrid design integrates LPC and NWC by redefining LPC’s quantization as a differentiable process, making the system training an end-to-end manner. The decoder of proposed system is with either one NWC (0.12 million parameters) in low to medium bitrate ranges (12 to 20 kbps) or two NWCs in the high bitrate (32 kbps). Although the decoding complexity is not yet as low as that of conventional speech codecs, it is significantly reduced from that of other neural speech coders, such as a WaveNet-based vocoder. For wide-band speech coding quality, our system yields comparable or superior performance to AMR-WB and Opus on TIMIT test utterances at low and medium bitrates. The proposed system can scale up to higher bitrates to achieve near transparent performance.

Index Terms:
Neural speech coding, waveform coding, representation learning, model complexity

I Introduction

Speech coding can be implemented as an encoder-decoder system, whose goal is to compress input speech signals into the compact bitstream (encoder) and then to reconstruct the original speech from the code with the least possible quality degradation. Speech coding facilitates telecommunication and saves data storage among many other applications. There is a typical trade-off a speech codec must handle: the more the system reduces the amount of bits per second (bitrate), the worse the perceptual similarity between the original and recovered signals is likely to be perceived. In addition, the speech coding systems are often required to maintain an affordable computational complexity when the hardware resource is at a premium.

For decades, speech coding has been intensively studied yielding various standardized codecs that can be categorized into two types: the vocoders and waveform codecs. A vocoder, also referred to as parametric speech coding, distills a set of physiologically salient features, such as the spectral envelope (equivalent to vocal tract responses including the contribution from mouth shape, tongue position and nasal cavity), fundamental frequencies, and gain (voicing level), from which the decoder synthesizes the speech. Typically, a vocoder operates at 33 kbps or below with high computational efficiency, but the synthesized speech quality is usually limited and does not scale up to higher bitrates [1][2][3]. On the other hand, a waveform codec aims to accurately reconstruct the input speech signal, which features up-to-transparent quality in a high bitrate range [4]. AMR-WB [5], for instance, can be seen as a hybrid waveform codec, because it employs speech modeling as in many other waveform codecs [6][7][8]. EVS [9], a recently standardized 3GPP voice and audio codec, has noticeably optimized frame error robustness, yielding a much-enhanced frame error concealment performance against than AMR-WB [10]. Similar to EVS, Opus, a waveform codec at its core, can also be applied to both speech and audio signals where it uses the LPC-based SILK algorithm for the speech-oriented model [11] and scales up to 510 kbps for transparent audio streaming and archiving.

Under the notion of unsupervised speech representation learning, deep neural network (DNN)-based codecs have revitalized the speech coding problem and provided different perspectives [12][13]. The major motivation of employing neural networks to speech coding is twofold: to fill the performance gap between vocoders and waveform codecs towards a near-transparent speech synthesis quality; to use its trainable encoder and learn latent representations which may benefit other DNN-implemented downstream applications, such as speech enhancement [14][15], speaker identification [16] and automatic speech recognition [17][18]. Having that, a neural codec can serve as a trainable acoustic unit integrated in future digital signal processing engines [13].

Recently proposed neural speech codecs have achieved high coding gain and reasonable quality by employing deep autoregressive models. The superior speech synthesis performance achieved in WaveNet-based models [19] has successfully transferred to neural speech coding systems, such as in [20], where WaveNet serves as a decoder synthesizing wideband speech samples from a conventional non-trainable encoder at 2.42.4 kbps. Although its reconstruction quality is comparable to waveform codecs at higher bitrates, the computational cost is significant due to the model size of over 2020 million parameters.

Meanwhile, VQ-VAE [12] integrates a trainable vector quantization scheme into the variational autoencoder (VAE) [21] for discrete speech representation learning. While the bitrate can be lowered by reducing the sampling rate 6464 times, the downside for VQ-VAE is that the prosody can be significantly altered. Although [22] provides a scheme to pass the pitch and timing information to the decoder as a remedy, it does not generalize to non-speech signals. More importantly, VQ-VAE as a vocoder does not address the complexity issue since it uses WaveNet as the decoder. Although these neural speech synthesis systems noticeably improve the speech quality at low bitrates, they are not feasible for real-time speech coding on the hardware with limited memory and bandwidth.

LPCNet [23] focuses on efficient neural speech coding via a WaveRNN [24] decoder by leveraging the traditional linear predictive coding (LPC) techniques. The input of the LPCNet is formed by 2020 parameters (1818 Bark scaled cepstral coefficients and 22 additional parameters for the pitch information) for every 1010 millisecond frame. All these parameters are extracted from the non-trainable encoder, and vector-quantized with a fixed codebook. As discussed previously, since LPCNet functions as a vocoder, the decoded speech quality is not considered transparent [1].

TABLE I: Categorical summary of recently proposed neural speech coding systems. ✔ means the system supports the feature while ✗  does not. ⚫ means it is not known.
WaveNet [20] VQ-VAE [22] LPCNet [23] Proposed
Transparent coding
Less than 1M parameters
Real-time processing
Encoder trainable

In this paper, we propose a novel neural speech coding system, with a lightweight design and scalable performance. First, we design a generic neural waveform codec with only 0.350.35 million parameters where 0.120.12 million parameters belong to the decoder. Compared to our previous models in [25][26] where the decoder has 0.230.23 million parameters, the current neural codec employs gated linear units to boost the gradient flow during model training and depthwise separable convolution to achieve further efficiency during decoding, as detailed in Sec. II. Based on this neural codec, our full system features two mechanisms to integrate speech production theory and residual coding techniques in Sec. III. Benefited from the residual-excited linear prediction (RELP) [27], we conduct LPC and apply the neural waveform codec to the excitation signal, which is illustrated in Sec.III-A. In this integration, a trainable quantizer bridges the encoding of linear spectral pairs and the corresponding LPC residual, making the speech coding pipeline end-to-end trainable. We also enable residual coding among neural waveform codecs to scale up the performance for high bitrates (Sec.III-B). In summary, the proposed system has following characteristics:

  • Scalability: Similar to LPCNet [23], the proposed system is compatible with conventional spectral envelope estimation techniques. However, ours operates at a much wider bitrate range with comparable or superior speech quality to standardized waveform codecs.

  • Compactness: The neural waveform codec in our system is with a much lower complexity than WaveNet [19] and VQ-VAE [12] based codecs. Our decoder contains only 0.120.12 million parameters which is 160×160\times more compact than a WaveNet counterpart. Our TensorFlow implementation's execution time to encode and decode a signal is only 42.44%42.44\% of its duration on a single-core CPU in the low-to-medium bitrates and 80.21% in the high bitrate, facilitating real-time communications.

  • Trainability: Our method is with a trainable encoder as in VQ-VAE, which can be integrated into other DNNs for acoustic signal processing. Besides, it is not constrained to speech, and can be generalized to audio coding with minimal effort as shown in [28].

TABLE I highlights the comparison to the other existing neural speech codecs.

This paper is an extension of the authors' previous conference presentations [25][26], where some initial ideas were already discussed. The new contributions presented this journal version are listed as follows:

  • Novel algorithmic enhancements: We propose new neural network architectures to form a new baseline autoencoder module and used it everywhere in our codecs. In our previous works, we have used a 1D convolutional neural network (CNN) that defines an autoencoder block with an identity shortcut as in the ResNet architecture [29]. While this architecture has been effective, in this journal paper, we propose to use the dilated gated linear units and depthwise separable convolution to reduce the kernel size without inducing any performance degradation. Consequently, our NWC is defined by 0.35M parameters, whose decoder part accounts for only 0.12M parameters. Compared to our previous models that are already small with only 0.45M parameters, the newly introduced reduction amounts to 22.2%22.2\%. If we only compare the decoder parts, it is a 47.8%47.8\% reduction. Although the proposed architecture is more compact than our previous models or the WaveNet-based codecs, since neural codecs' complexity is much larger than the traditional speech codecs, the additional model complexity reduction with no degradation of performance is promising. The architectural improvement are presented in Sec. II-A.

  • Extensive experimental validation: In our previous works, the experimental validation was to prove the initial concepts individually proposed each paper. In this time, we conduct an extensive and thorough experiments to provide the readers with a full view to the whole building-blocks of the neural speech coding. To this end, we define four candidate systems, from Model-I to IV, by incrementally adding new modules, such as LPC, the trainable LPC quantizer, and multiple concatenated neural autoencoders. The objective and subjective tests validate each of these additions in a full view (TABLE III and Fig. 7).

  • Additional analyses and ablation tests: We also provide detailed experimental validation for most of the claims made in the paper by designing and performing separate experiments, which were missing in the previous papers.

    • Sec. IV-D1 provides experimental verification that the proposed compact neural architecture does not induce performance loss.

    • Sec. IV-D2 presents a detailed analysis of the behavior of the cascaded autoendoers and the impact of different training strategies.

    • Sec. IV-F1 explores contribution of different loss terms in our training objective by performing ablation tests, and then proposes an optimal combination of hyperparameters.

    • Sec. IV-F2 also conducts an ablation test to empirically verify that the proposed trainable LPC quantization algorithm improves speech quality at the same bitrate.

    • Sec. IV-F3 and IV-F4 analyze the bit allocation behavior among the different submodules. Since the bit allocation strategy is decided by the learning algorithm, these analyses provide evidence that our models dynamically adapt to the characteristics of the signals given the limited bit budget.

    • Sec. IV-G presents additional analyses on computational complexity and execution time ratios to discuss the potential of the neural codecs in real-time applications.

    • Last but not least, in Sec. IV-G3 we discuss the implementation issues and the limitations of the proposed system in the context of real-world application scenarios.

Refer to caption
(a) The high-level structure of proposed neural waveform codec
Refer to caption
(b) Dilated gated linear unit (GLU)
Refer to caption
(c) Depthwise separable 1D convolution for upsampling
Figure 1: The proposed architecture for lightweight NWC.

II End-to-End Neural Waveform Codec (NWC)

The neural waveform codec (NWC), is an end-to-end autoencoder that forms the base of our proposed coding systems. NWC directly encodes the input waveform 𝒙T\bm{x}\in\mathbb{R}^{T} using a convolutional neural network (CNN) encoder enc(){\mathcal{F}}_{\text{enc}}(\cdot), i.e., 𝒉enc(𝒙){\bm{h}}\leftarrow{\mathcal{F}}_{\text{enc}}(\bm{x}). Then, the quantization process 𝒬(){\mathcal{Q}}(\cdot) converts the encoding into a bitstring 𝒉~N\tilde{\bm{h}}\in\mathbb{R}^{N}, which is followed by lossless data compression and bitstream transmission. On the receiver side, the decoder reconstructs the waveform as 𝒙𝒙^dec(𝒉~)\bm{x}\approx\hat{\bm{x}}\leftarrow{\mathcal{F}}_{\text{dec}}(\tilde{\bm{h}}). Fig. 1 (a) depicts NWC's overall system architecture. The structure is detailed in TABLE II. It serves as a basic component in the proposed speech coding system in Sec. III. In this section, we first introduce the architectural improvement that reduced our model's complexity. Next, we also introduce two strategies that compress the signals: feature map compression and trainable quantization.

II-A The Improved Architecture for NWC

We propose two different kinds of structural modification to reduce the model's overall complexity compared to other NWC models including our prior works [25, 26].

First, both encoder and decoder adopt gated linear units (GLU) [30]. We also define the GLU's convolution with dilation [31] to expand the receptive field in the time domain, which is a scheme that showed promising performance in speech enhancement [32]. Fig. 1 (b) shows our dilated GLU module. It first reduces the channel from 100 to 20 using a unit-width kernel. Then, two separate dilated convolution layers are applied to produce two feature maps, one of which goes through a sigmoid activation. Hence, the Hadamard product of the two feature maps can be seen as a gated version of the linear feature map. It is known that this gating mechanism in the middle boosts the gradient flow thanks to the linear path that does not involve the gradient vanishing issue. The final result is a mixture of the input and the last feature map, turning this block into a residual learning function as proposed in ResNet [29]. This kind of architecture shows superior performance as evidenced in [30, 33].

Our encoder reduces the data rate by using a ``downsampling" convolutional layer, whose ``stride" parameter is set to be 2. As a counterpart, the decoder's ``upsampling" layer makes up the loss. More details about this down and upsampling operations will be discussed in Sec. II-B. In this subsection, we focus on the actual module that performs the upmixing, where we introduced additional reduction in complexity. Out of various choices, we employ the depthwise separable convolution [34] to further save the computational cost (Fig. 1 (c)). For example, to transform a feature map of size 256×100256\times 100 (features, channels) into its upsampled version of size 512×50512\times 50, we first perform depthwise convolution using a c×1c\times 1 kernel. In our system c=9c=9. Since the depthwise convolution applies to each channel separately, we eventually need 100100 such kernels. It is a reduction of model complexity, because a normal convolution requires a kernel of size c×100×100c\times 100\times 100 (features, input channels, output channels), which is 100 time larger. In depthwise separable convolution, another 1×11\times 1 convolution follows to add more nonlinearity, for which we need a 1×100×1001\times 100\times 100 kernel. In our case, it is easy to show that c×100×100>c×100+100×100c\times 100\times 100>c\times 100+100\times 100 when the integer c>1c>1. For example, when c=9c=9, it is a reduction of about 88% of parameters.

Likewise, the proposed NWC is lightweight with only 0.35 million parameters, which is a reduction of 0.1 million parameters compared to our previous works [25][26]. The reduction comes from the streamlined upsampling operation implemented via the depthwise separable convolution. Eventually, the decoder accounts for 0.12 million parameters out of 0.35.

TABLE II: Architecture of the neural waveform codec: input and output tensors are shaped as (sample, channel), while the kernel is represented as (kernel size, in channel, out channel).
Layer Input shape Kernel shape Output shape
Encoder
Channel
Expansion
(512, 1) (55, 1, 100) (512, 100)
Gated
Linear
Unit
(512, 100)
(1, 100, 20) \rdelim]45mm[×\times2]
(15, 20, 20)
(15, 20, 20)
(9, 20, 100)
(512, 100)
Downsampling (512, 100) (9, 100, 100) (256, 100)
Gated
Linear
Unit
(256, 100)
(1, 100, 20) \rdelim]45mm[×\times2]
(15, 20, 20)
(15, 20, 20)
(9, 20, 100)
(256, 100)
Channel
Reduction
(256, 100) (9, 100, 1) (256, 1)
Decoder
Channel
Expansion
(256, 1) (9, 1, 100) (256, 100)
Gated
Linear
Unit
(256, 100)
(1, 100, 20) \rdelim]45mm[×\times2]
(15, 20, 20)
(15, 20, 20)
(9, 20, 100)
(256, 100)
Upsampling (256, 100)
(9, 100, 1)
(1, 100, 100)
(512, 50)
Gated
Linear
Unit
(512, 50)
(1, 50, 20) \rdelim]45mm[×\times2]
(15, 20, 20)
(15, 20, 20)
(9, 20, 50)
(512, 50)
Channel
Reduction
(512, 50) (55, 50, 1) (512, 1)

II-B Feature Map Compression

One way to compress the input signal in the proposed encoder architecture is to reduce the data rate. The CNN encoder function takes an input frame, 𝒙T\bm{x}\in\mathbb{R}^{T}, and converts it into a feature map 𝒉N\bm{h}\in\mathbb{R}^{N},

𝒉enc(𝒙),\bm{h}\leftarrow{\mathcal{F}}_{\text{enc}}(\bm{x}), (1)

which then goes through quantization, transmission, and decoding to recover the input as shown in Fig. 1 (a). During the encoding process, we introduce a downsampling operation, reducing the dimension of the code vector 𝒉\bm{h}. We employ a dedicated downsampling layer by setting up the stride value to be 22 during its convolution, reducing the data rate by 50%50\%, i.e., N=T/2N=T/2. Accordingly, the decoder needs a corresponding upsampling operation to recover the original sampling rate. We use subpixel CNN layer proposed in [35] to recover the original sampling rate. Concretely, the subpixel upsampling involves a feature transformation implemented in depthwise convolution, and a shuffle operation that interlaces features from two channels into a single channel, as shown in Eq. (2), where the input feature of the shuffle operation is shaped as (NN, 22) and the output is shaped as (2N2N, 11).

[h11,h21,h12,h22,,h1N,h2N]Upsampling([h11,h12,,h1N;h21,h22,,h2N]\begin{split}&[h_{11},h_{21},h_{12},h_{22},\ldots,h_{1N},h_{2N}]\\ &\leftarrow\text{Upsampling}([h_{11},h_{12},\ldots,h_{1N};h_{21},h_{22},\ldots,h_{2N}]\end{split} (2)

II-C The Trainable Quantizer for Bit Depth Reduction

The dimension-reduced feature map can be further compressed via bit depth reduction. Hence, the floating-point code 𝒉\bm{h} goes through quantization and entropy coding, which will finalize the bitrate based on the entropy of the code value distribution. Typically, a bit depth reduction procedure lowers the average amount of bits to represent each sample. In our case, we could employ a quantization process that assigns the output of the encoder to one of the pre-defined quantization bins. If there are 25=322^{5}=32 quantization bins, for example, a single-precision floating-point value's bit depth reduces from 32 to 5. In addition, various entropy coding techniques, such as Huffman coding, can be further employed to losslessly reduce the bit depth. While the quantization could be done in a traditional way, e.g., using Lloyds-Max quantization [36] after the neural codec is fully trained, we encompass the quantization step as a trainable part of the neural network as proposed in [37]. Consequently, we expect that the codec is aware of the quantization error, which the training procedure tries to reduce it. It is also convenient to control the bitrate by controlling the entropy of the code value distribution, which can be also done as a part of network training.

In NWC, the quantization process is represented as classification on each scalar value of the encoder output. Given a vector with KK centroids, 𝜷=[β1,β2,,βK]\boldsymbol{\beta}=[\beta_{1},\beta_{2},\cdots,\beta_{K}]^{\top}, the quantizer's goal is to assign each feature hnh_{n} to the closest centroid in terms of 2\ell_{2} distance, which is defined as follows:

𝑫=[h1β12h1βK2hNβ12hNβK2],\bm{D}=\begin{bmatrix}||h_{1}-\beta_{1}||_{2}&\cdots&||h_{1}-\beta_{K}||_{2}\\ \vdots&\ddots&\vdots\\ ||h_{N}-\beta_{1}||_{2}&\cdots&||h_{N}-\beta_{K}||_{2}\end{bmatrix}, (3)

where nn-th row in 𝑫\bm{D} is a vector of 2\ell_{2} distance between nn-th code value hnh_{n} to all KK quantization bins. Then, we employ the softmax function to turn each row of 𝑫\bm{D} into a KK-dimensional probabilistic assignment vector:

𝑨(soft)=[softmax(α𝑫1:)softmax(α𝑫2:)softmax(α𝑫N:)],\bm{A}^{\text{(soft)}}=\begin{bmatrix}\text{softmax}(-\alpha\bm{D}_{1:})\\ \text{softmax}(-\alpha\bm{D}_{2:})\\ \vdots\\ \text{softmax}(-\alpha\bm{D}_{N:})\\ \end{bmatrix}, (4)

where we turn the distance into a similarity metric by multiplying a negative number α-\alpha, such that the shortest distance is converted to the largest probability. In our implementation, 𝜷\boldsymbol{\beta} is initialized as a vector of K=32K=32 uniformly spaced numbers within the interval of [1,1][-1,1]. As for α\alpha, we begin with a large enough number 300300. Both α\alpha and β\beta are trainable parameters to optimize the quantization process.

Note that Eq. (4) yields a soft assignment matrix 𝑨(soft)N×K\bm{A}^{\text{(soft)}}\in\mathbb{R}^{N\times K}. In practice, though, the quantization process must perform a hard assignment, so each code value hnh_{n} is replaced by an integer index to the closest centroids: zn{1,2,,K}z_{n}\in\{1,2,\cdots,K\}, which is represented by log2K\lceil\log_{2}K\rceil bits as the quantization result. The hard kernel assignment matrix 𝑨(hard)\bm{A}^{\text{(hard)}}, where each row is a one-hot vector, can be induced by turning on the maximum element of 𝑨(soft)\bm{A}^{\text{(soft)}} while suppressing the non-maximum:

𝑨nk(hard)={1 if argmaxj{1,2,,K}𝑨nj(soft)=k0otherwise\bm{A}^{\text{(hard)}}_{nk}=\left\{\begin{array}[]{rl}1&\text{ if }\operatornamewithlimits{\arg\max}_{j\in\{1,2,\cdots,K\}}\bm{A}^{\text{(soft)}}_{nj}=k\\ 0&\text{otherwise}\end{array}\right. (5)

On the decoder side, 𝒉~=𝑨(hard)𝜷\tilde{\bm{h}}=\bm{A}^{\text{(hard)}}\bm{\beta} recovers 𝒉\bm{h}.

Algorithm 1 Soft-to-hard quantization during inference, 𝒬(𝒉,α,𝜷)\mathcal{Q}(\bm{h},\alpha,\boldsymbol{\beta})
1:  Input: the code, e.g., the encoder output, 𝒉=enc(𝒙)\bm{h}=\mathcal{F}_{\text{enc}}(\bm{x})          the trained softmax scaling factor, α\alpha          the trained centroid vector, 𝜷K\boldsymbol{\beta}\in\mathbb{R}^{K}
2:  Output: the quantized code, 𝒉^\hat{\bm{h}} (training) or 𝒉~\tilde{\bm{h}} (testing)
3:  Compute the dissimilarity matrix: 𝑫nk2(hn||βk)\bm{D}_{nk}\leftarrow\ell_{2}(h_{n}||\beta_{k})
4:  Softmax conversion: 𝑨n:(soft)Softmax(α𝑫n:)\bm{A}^{\text{(soft)}}_{n:}\leftarrow\text{Softmax}(-\alpha\bm{D}_{n:})
5:  if Training then
6:     Soft quantization: 𝒉^𝑨(soft)𝜷\hat{\bm{h}}\leftarrow\bm{A}^{\text{(soft)}}\boldsymbol{\beta}
7:  else if Testing then
8:     Hard quantization: 𝒉~𝑨(hard)𝜷\tilde{\bm{h}}\leftarrow\bm{A}^{\text{(hard)}}\boldsymbol{\beta}
9:  end if

Since argmax\arg\max operation in Eq. (5) is not differentiable, a soft-to-hard scheme is proposed in [37], where 𝑨(hard)\bm{A}^{\text{(hard)}} is used only at test time. During backpropagation for training, the soft classification mode is enabled with 𝑨(soft)\bm{A}^{\text{(soft)}} so as not to block the gradient flow. In other words, 𝒉^=𝑨(soft)𝜷\hat{\bm{h}}=\bm{A}^{\text{(soft)}}\bm{\beta} represents each encoder output with a linear combination of all quantization bins. The process is summarized in Algorithm 1. Although this soft quantization process is differentiable and desirable during training, the discrepancy between 𝑨(soft)\bm{A}^{\text{(soft)}} and 𝑨(hard)\bm{A}^{\text{(hard)}} creates higher error during the test time, requiring a mechanism to reduce the discrepancy as in the following section.

II-C1 Soft-to-hard quantization penalty

Although the limit of 𝑨(soft)\bm{A}^{\text{(soft)}} is 𝑨(hard)\bm{A}^{\text{(hard)}} as α\alpha approaches \infty, the change of α\alpha should be gradual to allow gradient flows in the initial phase of training. We control the hardness of 𝑨(soft)\bm{A}^{\text{(soft)}} using the soft-to-hard quantization loss derived from [38]:

Q=1Nn,k𝑨nk(soft),{\mathcal{L}}_{\text{Q}}=\frac{1}{N}\sum_{n,k}\sqrt{\bm{A}^{\text{(soft)}}_{nk}}, (6)

whose minimum, 1, is achieved when 𝑨n:(soft)\bm{A}^{\text{(soft)}}_{n:} is a one-hot vector for all nn. Conversely, when 𝑨nk(soft)=1/K\bm{A}^{\text{(soft)}}_{nk}=1/K, the loss is maximum. Hence, by minimizing this soft-to-hard quantization penalty term, we can regularize the model to have harder 𝑨(soft)\bm{A}^{\text{(soft)}} values by updating α\alpha and the other model parameters accordingly. As a result, the test time quantization loss will be reasonably small when 𝑨(soft)\bm{A}^{\text{(soft)}} is replaced by 𝑨(hard)\bm{A}^{\text{(hard)}}.

II-C2 Bitrate calculation and entropy control

The bitrate is calculated as a product of the number of code values per second and the average bit depth for each code. The former is defined by the dimension of the code vector NN multiplied by the number of frames per second, FTo\frac{F}{T-o}, where TT, oo, and FF are the input frame size, overlap size, and the original sampling rate, respectively. If we denote the average bit depths per sample by a function g(h~n)g(\tilde{h}_{n}), the bitrate can be computed as in Eq. (7),

bitrate=g(h~n)NF/(To).\text{bitrate}=g(\tilde{h}_{n})NF/(T-o). (7)

When F=16,000F=16,000, T=512T=512, o=32o=32, and N=256N=256 after downsampling, for example, there are about 8,5338,533 samples per second. If g(h~n)=3g(\tilde{h}_{n})=3 bits, the bitrate is estimated as 25.625.6 kbps. In contrast, the uncompressed bitrate is 256256 kbps with N=T=512N=T=512, o=0o=0, and g(xt)=16g(x_{t})=16 bits for each sample.

We alter the entropy of 𝜷\bm{\beta} to adjust the codec's bitrate, since the entropy serves as the lower bound of g(h~n)g(\tilde{h}_{n}) based on Shannon's entropy theory. The entropy, (𝜷)\mathcal{H}(\bm{\beta}), is estimated with the sample distribution,

(𝜷)k=1Kp(βk)log2p(βk),\mathcal{H}(\bm{\beta})\approx-\sum_{k=1}^{K}p(\beta_{k})\log_{2}p(\beta_{k}), (8)

where p(βk)=1Nn𝑨nk(hard)p(\beta_{k})=\frac{1}{N}\sum_{n}\bm{A}^{\text{(hard)}}_{nk} is the relative frequency of the kk-th centroid being chosen during quantization. To navigate the model training towards the target bitrate, (𝜷)\mathcal{H}(\bm{\beta}) defined in Eq. (8) is included in the loss function as a regularizer: if the current bitrate is higher than desired, the optimization process will increase the blending weight of the regularizer to strengthen the regularization effect, which consequently lowers the entropy, and vice versa. More details on the training target and hyperparameter setting are discussed in Sec. IV-B.

III NWC-based Speech Coding Systems

By having NWC introduced in Sec. II as the basic module, we propose two different extension mechanisms to improve the codec's performance in a wider range of bitrates, without increasing the model complexity significantly. First, in Sec. III-A, we propose a neural network compatible LPC module where the trainable soft-to-hard quantization is applied to the LPC coefficients. With the LPC module followed by an NWC module, we achieve a win-win strategy that fuses the traditional DSP technique and the modern deep learning model [26]. In addition to integrating LPC, our neural codec conducts multistage residual coding [25] by cascading residuals among multiple NWC modules (Sec. III-B). The proposed CMRL system relays residual signals among the series of NWCs to scale up the coding performance at high bitrates.

III-A Trainable LPC Analyzer

LPC has been widely used to facilitate speech compression and sysnthesis, where source-filter model ``explains out" the envelope of a speech spectrum, leaving a low-entropy residual signal [39]. Similarly, LPC serves as a pre-processor in our system before its residual signal being compressed by NWC as we will see in Sec. III-B. In this subsection, we redesign the LPC coefficient quantization process as a trainable module. We introduce collaborative quantization (CQ) to jointly optimize the LPC analyzer and NWCs as a residual coder.

III-A1 Speech resonance modeling

In the speech production process, the source as wide-band excitation signals go through the vocal tract tube. The shape-dependent resonances of the vocal tract filter the excitations before it being transformed to speech signals [40]. In speech coding, the ``vocal tract response" is often modeled as an all-pole filter [41]. Having that, the tt-th sample xtx_{t} can be approximated by a autoregressive model using MM previous samples,

xt=k=1Mlkxtk+et,x_{t}=\sum\limits_{k=1}^{M}l_{k}x_{t-k}+e_{t}, (9)

where the estimation error ete_{t} represents the LPC residual, and lkl_{k} denotes the filter coefficients. Typically, lk{l_{k}} can be efficiently estimated via Levinson-Durbin algorithm [42], and are to be quantized before LPC residual is calculated, i.e., ete_{t} encompasses the quantization error. The LPC residual ete_{t} serves as input to the NWC module, which works as explained in Sec. II, but on 𝒆\bm{e} rather than 𝒙\bm{x}. Hence, how LPC coefficients are quantized determines NWC's input, the LPC residual.

III-A2 Collaborative quantization

The conventional LPC coefficient quantization process is standardized in ITU-T G.722.2 (AMR-WB) [43]: 2.42.4k bits are assigned to represent the LPC coefficient per second though multistage vector quantization (MSVQ) [44] in a classic LPC analyzer. Once again, we employ the soft-to-hard quantizer as illustrated in Sec. II to make the quantization and bit allocation steps in the LPC analyzer trainable and communicatable with the neural codec.

Refer to caption
(a) The trainable LPC analyzer where the linear spectral pairs are quantized by our trainable soft-to-hard quantizer (in the dotted box).
Refer to caption
(b) Cross-frame windowing
Refer to caption
(c) Sub-frame windowing
Refer to caption
(d) Synthesis windowing
Figure 2: The signal flow chart for LPC analyzer (a) and windowing schemes in LPC (b)-(d).

We compute the LPC coefficients as in [5], first by applying high-pass filtering followed by pre-emphasizing (Fig. 2 (a)). When calculating LPC coefficients, the window in Fig. 2 (b) is used. The window is symmetric with the left and right 25% parts being tapered by a 512-point Hann window. After representing the 16 LPC coefficients in linear spectral pairs (LSP) [45], we quantize it using the soft-to-hard quantization scheme. Then, the sub-frame window in Fig. 2 (c) is applied to calculate LPC residual, which assures a more accurate residual calculation. The frame that covers samples [256:768], for instance, is decomposed into 7 sub-frames to calculate LPC residuals separately. Each 128-point Hann window in Fig. 2 (c) is with 50% overlap, except for the first and last window. They altogether form a constant overlap-add operation. Finally, after the synthesis using the reconstructed residual signal and corresponding LPC coefficients, the window in Fig. 2 (d) tapers both ends of the synthesized signal, covering 512 samples with 32 overlapping samples between adjacent windows.

As an intuitive example, given the samples [1:1024] as the input, after the LPC analysis, neural residual coding, and LPC synthesis, samples [257:768] are decoded; the next input frame is [481:1504] (the dotted window in Fig. 2 (b)), whose decoded samples are within [737:1248]. The overlap-add operation is applied to the final decoded samples [737:768] (Fig. 2 (d)).

During this process, the calculated LPC coefficients are quantized using Algorithm 1, where the code vector is with 16 dimensions, i.e., 𝒉16\bm{h}\in\mathbb{R}^{16}. The number of kernels is set to be K=28=256K=2^{8}=256. Note that the soft assignment matrix for the LPC quantization, 𝑨(LPC)\bm{A}^{(\text{LPC})}, is also involved in the loss function to regularize the bitrate.

We investigate the impact of the trainable LPC quantization in collaboration with the rest of the NWC modules in Sec. IV.

III-B Cross-Module Residual Learning (CMRL)

To achieve scalable coding performance towards transparency at high bitrates, we propose cross-module residual learning (CMRL) to conduct bit allocation among multiple neural codecs in a cascaded manner. CMRL can be regarded as a natural extension of what is described in Sec. III-A, where the LPC as a codec conducts the first round of coding by only modeling the spectral envelope. It leaves the residual signal for a subsequent NWC to be further compressed. With CMRL, we employ the concept of residual coding to cascade more NWCs. We also present a dual-phase training scheme to effectively train the CMRL model.

CMRL's scalability comes from its residual coding concept that enables a concatenation of multiple autoencoding modules. We define the residual signal recursively: ii-th codec takes the residual of its predecessor as input, and the ii-th reconstruction creates another residual for the next round, and so on. Hence, we have

𝒙^(i)(i)(𝒙(i)),𝒙(i)𝒙(i1)𝒙^(i1),𝒙(1)𝒙,\hat{\bm{x}}^{(i)}\leftarrow{\mathcal{F}}^{(i)}(\bm{x}^{(i)}),~{}\bm{x}^{(i)}\leftarrow\bm{x}^{(i-1)}-\hat{\bm{x}}^{(i-1)},~{}\bm{x}^{(1)}\leftarrow\bm{x}, (10)

where 𝒙^(i)\hat{\bm{x}}^{(i)} stands for the reconstruction of the ii-th input using the ii-th coding module (i)(){\mathcal{F}}^{(i)}(\cdot), while the input to the first codec is defined by the raw input frame 𝒙\bm{x}. If we expand the recursion, we arrive at the non-recursive definition of 𝒙(i)\bm{x}^{(i)},

𝒙(i)=𝒙j=1i1𝒙^(j),\displaystyle\bm{x}^{(i)}=\bm{x}-\sum\limits_{j=1}^{i-1}\hat{\bm{x}}^{(j)}, (11)

which means the input to ii-th model is the residual of the sum of all preceding i1i-1 codecs' decoded signals. It ensures the additivity of the entire system: adding more modules keeps improving the reconstruction quality. Hence, CMRL can scale up to high bitrates at the cost of increased model complexity.

CMRL is optimized in two phases. During Phase-I training, we sequentially train each codec from the first to the last one using a module-specific residual reconstruction goal,

(𝒙(i)||𝒙^(i)).{\mathcal{E}}(\bm{x}^{(i)}||\hat{\bm{x}}^{(i)}). (12)

The purpose for Phase-I training is to get parameters for each codec properly initialized. Then, Phase-II finetunes all trainable parameters of the concatenated modules to minimize the global reconstruction loss,

(𝒙||i=1N𝒙^(i)).{\mathcal{E}}\left(\bm{x}\bigg{|}\bigg{|}\sum_{i=1}^{N}\hat{\bm{x}}^{(i)}\right). (13)

The reconstruction loss measures the waveform discrepancy in both time and frequency domains. Quantization penalty and entropy control are introduced as regularizers. Sec. IV-B details the definition of the training target and ablation tests on how to find the optimal blending weights between the loss terms.

Refer to caption
Figure 3: The flow diagram of the test-time inference.

III-C Signal Flow during Inference

Fig. 3 shows the full system signal flow with NN sub-codecs, having an LPC module as the first one. On the transmitter side the LPC analyzer first processes the input frame 𝒙\bm{x} of 1024 samples and computes 16 coefficients, 𝒉(1)\bm{h}^{(1)}, as well as 512 residual samples 𝒙(2)\bm{x}^{(2)} at the center of the frame. Then, the residual signal goes through the N1N-1 NWCs in sequentially. Note that the transmission process's primary job is to produce a quantized bitstring 𝒉~(i)\tilde{\bm{h}}^{(i)} from LPC and each NWC. To this end, NWC's decoder part must also run to compute the residual signal and relay it to the next NWC module. The bitstring is generated as a concatenation of all encoder outputs: 𝒉~=[𝒉~(1);𝒉~(2);;𝒉~(N)]\tilde{\bm{h}}=\big{[}{\tilde{\bm{h}}^{(1)}};{\tilde{\bm{h}}^{(2)}};\cdots;{\tilde{\bm{h}}^{(N)}}\big{]}. Once the bitstring is available on the receiver side, all NWC decoders run to reconstruct the LPC residual signal, i.e., 𝒙^(2)i=2Ndec(i)(𝒉~(i))\hat{\bm{x}}^{(2)}\approx\sum_{i=2}^{N}{\mathcal{F}}_{\text{dec}}^{(i)}(\tilde{\bm{h}}^{(i)}). Then it is used as input of the LPC synthesizer, along with the LPC coefficients.

IV Evaluation

In this section, we examine the proposed neural speech coding model presented in Sec.II and III. The evaluation critera include both objective measures such as PESQ [46] and signal-to-noise ratio (SNR) and subjective scores from MUSHRA listening tests [47]. In addition, we conduct ablation analysis to provide a detailed comparison between various loss terms and bit allocation schemes. Finally, we report the system delay and execution time under four hardware specifications.

IV-A Data Processing

The training dataset is created from 300 speakers randomly selected from the TIMIT corpus [48] with no gender preference. Each speaker contributes 10 utterances totaling 2.6 hour-long training set, which is a reasonable size due to our compact design. The same scheme is adopted when creating the validation dataset and test dataset with 50 speakers, respectively. All three datasets are mutually exclusive with the sample rate of 1616 kHz. All neural codecs in this work are trained and tested via the same set of data for a fair comparison. We normalize each utterance to have a unit variance, then divided by the global maximum amplitude, before being framed into segments with the size of 512 samples. On the receiver side, we conduct overlap-and-add after the synthesis of the frames, where a 32-sample Hann window is applied to the overlapping region of the same size.

With the LPC codec, we apply high-pass filtering defined in the z-space, 𝒢hp(z)=0.9895021.979004z10.989502z211.97888210.9791262\mathcal{G}_{\text{hp}}(z)\!=\frac{0.989502-1.979004z^{-1}0.989502z^{-2}}{1-1.978882^{-1}0.979126^{-2}}, to the normalized waveform. A pre-emphasis filter, 𝒢preemp(z)=10.68z1\mathcal{G}_{\text{preemp}}(z)\!=1\!-\!0.68z^{-1}, follows to boost the high frequencies.

IV-B Training Targets and Hyperparameters

The loss function is defined as

=λMSEt=1T(xtx^t)2+λmelb=14f=1Fb(yf(b)y^f(b))2+λQQ+λent(𝜷)\begin{split}\mathcal{L}&=\lambda_{\text{MSE}}\sum_{t=1}^{T}(x_{t}-\hat{x}_{t})^{2}+\lambda_{\text{mel}}\sum_{b=1}^{4}\sum_{f=1}^{F_{b}}\Big{(}y_{f}^{(b)}-\hat{y}_{f}^{(b)}\Big{)}^{2}\\ &+\lambda_{\text{Q}}{\mathcal{L}}_{\text{Q}}+{\color[rgb]{0,0,0}{\lambda_{\text{ent}}}}\mathcal{H}(\bm{\beta})\\ \end{split} (14)

where the first term measures the mean squared error (MSE) between the raw waveform samples and their reconstruction. Ideally, if the model complexity and the bitrate is sufficiently large, an accurate reconstruction is feasible by using MSE as the only loss function. Otherwise, the result is usually sub-optimal due to the lack of bits: coupled with the MSE loss, the decoded signals tend to contain broadband artifact. The second term supplements the MSE loss and helps suppress this kind of artifact. To this end, we follow the common steps to conduct mel-scaled filter bank analysis, which results in a mel spectrum 𝒚\bm{y} that has a higher resolution in the low frequencies than in the high frequencies. The filter bank size defines the granularity level of the comparison. Following [38], we conduct a coarse-to-fine filter bank analysis by setting four filter bank sizes, F1=8,F2=16,F3=32,F4=128F_{1}=8,F_{2}=16,F_{3}=32,F_{4}=128 as shown in Fig. 4, which result in four kinds of resolutions for mel spectra 𝒚(b)\bm{y}^{(b)} indexed by b{1,2,3,4}b\in\{1,2,3,4\}.

Refer to caption
Figure 4: The coarse-to-fine filter bank analysis in the mel scale.

All models are trained on Adam optimizer with default learning rate adaptation rates [49]. The batch size is fixed with 128128 frames . The initial learning rate is 2×1032\times 10^{-3} for the first neural codec. With CMRL, the learning rate for the successive neural codecs is 2×1042\times 10^{-4}. Finetuning of all those models is with a smaller learning rate 2×1052\times 10^{-5}. All models are sufficiently trained until the validation loss converges after being exposed to about 5×1055\times 10^{5} batches. These hyperparameters were chosen based on validation.

The blending weights in the loss function in Eq. (14) are also selected based on the validation performance. Empirically, the ratio between the time-domain loss and mel-scaled frequency loss affects the trade-off between the SNR and perceptual quality of decoded signals. If the time-domain loss dominates the optimization process, the model compresses each sub-band with an equal effort. In that case, the artifact will be audible unless the SNR reaches a rather high level (over 3030 dB) which entails a high bitrate and model complexity. On the other hand, if only the mel-scaled frequency loss is in place, the reconstruction quality in the high frequency will degrade. The impact of these blending weights for these two loss terms is detailed in Sec. IV-F via an ablation analysis.

The weights for the quantization regularizer λQ\lambda_{\text{Q}} and entropy regularizer λent\lambda_{\text{ent}} are initially set to be 0.50.5 and 0.00.0, respectively. As for λent\lambda_{\text{ent}}, we alter it after every epoch by 0.0150.015: if the current model's bitrate is higher than the target bitrate, λent\lambda_{\text{ent}} increases to penalize the model's entropy more; otherwise, λent\lambda_{\text{ent}} decreases to boost the entropy. Note that we omit the module index ii in Eq. (14), so the meaning of x^t\hat{x}_{t} depends on the context: either the module-specific reconstruction as in Eq. (12) or the sum of all recovered residual signals for Phase-II finetuning as in Eq. (13). Similarly, Q{\mathcal{L}}_{\text{Q}} and 𝜷{\mathcal{H}}_{\bm{\beta}} can encompass all modules' quantization and entropy losses including LPC's for Phase-II. We delay the introduction of the quantization and entropy loss until the fifth epoch.

IV-C Bitrate Modes and Competing Models

We consider three bitrates, 1212, 2020, and 3232 kbps, to validate models' performance in a range of use cases. We evaluate following different versions of neural speech coding systems:

  • Model-I: The NWC baseline (Sec. II).

  • Model-II: Another baseline that combines the legacy LPC and an NWC module for residual coding.

  • Model-III: A trainable LPC quantization module followed by an NWC and finetuning (Sec. III-A);

  • Model-IV: Similar to Model-III but with two NWC modules: the full-capacity CMRL implementation (Sec.III-B). It is tested cover the high bitrate case, 3232 kbps.

Regarding the standard codecs, AMR-WB [5] and Opus [50] are considered for comparison. AMR-WB, as an ITU standard speech codec, operates in nine different modes covering a bitrate range from 6.6 kbps to 23.85 kbps, providing excellent speech quality with a bitrate as low as 12.65 kbps in wideband mode. As a more recent codec, Opus shows the state-of-the-art performance in most bitrates up to 510 kbps for stereo audio coding, except for the very low bitrate range.

We first compare all models with respect to the objective measures, while being aware that they are not consistent with the subjective quality. Hence, we also evaluate these codecs in two rounds of MUSHRA subjective listening tests: the neural codecs are compared in the first round, whose winner is compared with other standard codecs in the second round.

TABLE III: Objective measurements for neural codec comparison under three bitrate cases.
Bitrate SNR (dB) PESQ-WB
(kbps) Model-I Model-II Model-III Model-IV AMR-WB Opus Model-I Model-II Model-III Model-IV AMR-WB Opus
\sim12 12.37 10.69 10.85 11.60 9.63 3.67 3.45 3.60 3.92 3.93
\sim20 16.87 10.73 13.65 13.14 9.46 4.37 3.95 4.01 4.18 4.37
\sim32 20.24 11.84 14.46 17.11 17.66 4.42 4.15 4.18 4.35 4.38

IV-D Objective Measurements

IV-D1 The compact NWC module and its performance

Compared to our previous models in [25][26] that use 0.450.45 million parameters, the newly proposed NWC in this work only has 0.350.35 million parameters. It is also a significant reduction from the other compact neural waveform codec [38] with 1.61.6 million parameters. As introduced in Sec. II the model size reduction is achieved via the GLU [33] and depthwise separable convolution for upsampling [34]. In our first experiment, we show that the objective measures stay the same. Fig. 5 compares the NWC modules before and after the structural modification proposed in Sec. II in terms of (a) signal-to-noise ratio (SNR) and (b) PESQ-WB [46]. We can see that the newly proposed model with 0.350.35M parameters is comparable to the larger model. Therefore, it justifies its use as the basic module in a range of models from Model-I to IV.

Refer to caption
(a) The validation SNR curve during training
Refer to caption
(b) The validation PESQ curve during training
Figure 5: Speech reconstruction performance stays almost the same when the model size decreases from 0.45 to 0.35 million parameters with the help from the structural modification.
Refer to caption
(a) Scalability with respect to SNR
Refer to caption
(b) Scalability with respect to PESQ
Figure 6: In CMRL, performance leaps when the new neural codec is added for residual cascading. In these CMRL models LPC is not used.

IV-D2 The impact of CMRL's residual coding

To validate the merit of CMRL's residual coding concept, we scale up the CMRL model by incrementally adding more NWC modules up to five. In Fig. 6, both SNR and PESQ values keep increasing when CMRL keeps adding a new NWC module. There are two noticeable points in these graphs. First, the greedy module-wise pretraining is important for the performance: whenever a new model is added, it is pretrained to minimize the module specific loss Eq. (12) first (Phase-I), then the global loss Eq. (13), subsequently (Phase-II). A model that does not perform Phase-II (thick gray line) stagnates no matter how many NWCs are added. Second, we also train a very large NWC model with the same amount of parameters as CMRL with five NWCs combined (grey dash). It turns out the equally large model fails to scale up due to its single integrated architecture. While we eventually decide to use only up to two NWCs for speech coding for our highest bitrate case, 3232 kbps, we may keep adding NWCs to CMRL to meet the case of higher bitrates for non-speech audio coding.

IV-D3 Overall objective comparison of all competing models

TABLE III reports SNR and PESQ-WB from all competing systems. AMR-WB in the low-range bitrate setting operates at 12.65 kbps and 23.05 kbps for the mid-range. Among neural speech coding systems, Model-I, as a single autoencoder model, outperforms others in all three bitrate setups in terms of SNR and PESQ-WB. It is also comparable to AMR-WB and Opus, except for the low bitrate case where Opus achieves the highest PESQ score. One explanation is that Model-I is highly optimized for the objective loss during training, although it does not necessarily mean that the higher objective score leads to a better subjective quality as presented in Sec. IV-E. It is also observed that with CQ, Model-III gains slightly higher SNR and PESQ scores compared to Model-II, which uses the legacy LPC. Finally, the performance scales up significantly when Model-IV starts to employ two NWCs on top of LPC, which is our proposed full neural speech coding setup. Aside from objective measure comparison, to further evaluate the quality of proposed codec, we discuss the subjective test in the next section.

Refer to caption
(a) Neural waveform codecs comparison.
Refer to caption
(b) Comparison between proposed methods and standard codecs.
Figure 7: MUSHRA subjective listening test results.

IV-E Subjective Test

We conduct two rounds of MUSHRA tests: (a) to select the best one out of the proposed models (from Model-I to IV) (b) to compare it with the standard codecs, i.e., AMR-WB and Opus. Each round covers three different bitrate ranges, totaling six MUSHRA sessions. A session consists of ten trials, for which ten gender-balanced test signals are randomly selected. Each trial has one low-pass filtered signal serving as the anchor (with a cutoff frequency at 4kHz), the hidden reference, as well as signals decoded from competing systems. We recruit ten participants who are audio experts with prior experiences in speech/audio quality evaluation. The subjective scores are rendered in Fig. 7 as boxplots. Each box ranges from the 2525 to 7575 percentile with a 95%95\% confidence interval. The mean and median are presented as the green dotted line and pink hard line, respectively. Outliers are represented in circles.

Refer to caption
Figure 8: The ablation analysis on CQ.
Refer to caption
Figure 9: The frame-wise bit allocation analysis.

IV-E1 Comparison among the proposed neural codecs

In Fig. 7 (a) we see that Model-III's produces decoding results that are much more preferred than both Model-I and Model-II, which are a pure end-to-end model and with the non-trainable legacy LPC module, respectively. The advantage is more noticeable in lower bitrates. It is contradictory to the objective scores reported in TABLE III where Model-I often achieves the highest scores. Compared to the deterministic quantization component in the legacy LPC in Model-II, LPC module and NWC in Model-III are jointly trained for a frame-wise independent bit allocation, so as to maximize the coding efficiency. We also note that Model-III's performance stagnates in the high bitrate experiments, suggesting its poor scalability. To this end, for the high bitrate experiment, we additionally test Model-IV with two NWC residual coding modules instead of just one. Model-IV outperformed both Model-II and Model-III by a large margin, showcasing a near-transparent quality.

IV-E2 Comparison with standardized codecs

Fig. 7 (b) shows that, our Model-II is on par with AMR-WB for the low-range bitrate case, while outperforming Opus which tends to lose high frequency components. In the medium-range, Model-II at 19.219.2 kbps is comparable to Opus at 20.020.0 kbps and AMR-WB at 23.0523.05 kbps. In the high bitrate range, our Model-IV outperforms Opus that operates 3232 and 2424 kbps, while AMR-WB is omitted as it does not support those high bitrates.

All MUSHRA sessions are available online, along with demo samples and source codes111https://saige.sice.indiana.edu/research-projects/neural-audio-coding.

TABLE IV: Ablation analysis on blending weights.

(a) Model-I (a neural codec only)

Blending Ratio (MSE : mel) Decoded SNR (dB) Decoded PESQ
1:01:0 18.12 3.67
0:10:1 0.16 4.23
1:11:1 6.23 4.31
10:110:1 16.88 4.37

(b) Model-III (a neural codec with a collaboratively trained LPC)

Blending Ratio Residual SNR Decoded SNR Decoded PESQ
(MES : mel) (dB) (dB)
1:01:0 9.73 14.25 3.84
0:10:1 1.79 17.23 4.02
1:11:1 7.11 17.82 4.08
10:110:1 8.26 17.55 4.01

IV-F Ablation Analysis

In this section, we perform some ablation analyses to justify our choices that led to CQ and CMRL's superior subjective test results. We investigate how different blending ratios between loss terms can alter the performance. We will also explore the optimal bit allocation strategy among coding modules.

IV-F1 Blending weights for the loss terms

Out of the two reconstruction loss terms, MSE serves as the main loss for the end-to-end NWC system, while the mel-scaled loss prioritizes certain frequency bands over the others. We perform ablation analysis on four blending ratio settings to analyze their effect on decoded speech's objective quality. We consider both system configurations: one with only the neural waveform codec (TABLE IV (a)) and the other one with both the neural waveform codec and collaboratively trained LPC module (TABLE IV (b). For TABLE IV (a), the target entropy for each sample in the neural codec is of 2.52.5-bit, corresponding to a bitrate of 512321600010242562.521.3\frac{512-32}{16000*1024}*256*2.5\approx 21.3 kbps. The SNR reaches the highest when there is only the MSE term, while the PESQ score becomes the lowest. By only keeping the mel-scaled loss term, the PESQ score is decent (4.23), yet with a poor waveform reconstruction as suggested by the SNR value (0.16 dB). For TABLE IV (b), the target entropy for each LPC coefficient is of 55-bit or 2.62.6 kbps, and 3.53.5-bit for each LPC residual sample or 29.929.9 kbps. Similarly, even with the input of the neural codec being the LPC residuals, MSE alone yields the highest SNR for the reconstruction of the LPC residual, but it does not benefit the final synthesized signal even in terms of SNR. Note that we choose 128 quantization centroids for the high bitrate case, which is different from that of Model-III in TABLE III where 32 quantization centroids are employed. For consistency's sake, we choose the blending ratio of 10:110:1, which shows reasonably well numbers in all proposed models.

IV-F2 CQ's impact on the speech quality

We compare the PESQ values of the decoded signals from Model II and III. Since Model-III shares the same architecture with Model-II except for the CQ training strategy, the comparison is to verify that CQ can effectively allocate bits to the LPC and NWC modules. Fig. 9 shows that the total entropy of the two models are under the control regardless of the use of CQ mechanism. However, we can see that Model-III with CQ achieves higher PESQ during and after the control of the entropy, showcasing that the CQ approach benefits the codec's performance.

IV-F3 Bit allocation between the LPC and NWC modules

Since the proposed CQ method is capable of assigning different bits to the LPC and NWC modules dynamically, i.e., in a frame-by-frame manner, we analyze its impact in more detail. In the mid-range bitrate setting, Fig. 9 shows the amount of bits assigned to both modules per frame (b/f). First of all, we observe that the dynamic bit allocation scheme indeed adjusts the LPC and NWC bitrates over time. Because of the CQ-enabled dynamic bit allocation, our system is able to compress silent frames more efficiently: by allocating just a few more bits to LPC, it saves a lot more bits from the NWC module for residual coding, as shown in the crimson-colored boxed areas in Fig. 9. However, it still requires a significant amount of bits to even represent those near-silent frames, which can be further optimized by voice activity detection. Finally, it appears that NWC is less efficient for fricatives (e.g., f and S) and affricates (e.g., tS). TABLE V shows the overall bit allocation among different modules. In the low bitrate case, it is worth noting that CQ uses 58 b/f or 1.931.93 kbps, differently from AMR-WB's standard, 2.42.4 kbps.

IV-F4 Bit allocation between the two NWC modules in Model-IV

To find the optimal bit allocation between two NWC modules, we first conduct an ablation analysis on 33 different bit allocation choices. In Fig. 10, both the SNR and PESQ scores degrade when the second NWC uses 33.3% more bits than the first one. Among these 33 choices, the highest PESQ score is obtained when the first NWC module uses 33.3% more bits. In practice, the bit allocation is automatically determined during the optimization process. In TABLE V, for example, the bit ratio between two NWC modules of Model-IV in the high bitrate case is about 486:38455.9:44.1486:384\approx 55.9:44.1, in accord with the observation from the ablation analysis that the first module should use more bits.

TABLE V: Bit allocation among coding components.
Bitrate Modes LPC Coefficients LPC Residual Total
(kbps) (bits / frame) (bits / frame) (bits / frame)
11.77\sim 11.77 58 295 353
19.20\sim 19.20 74 502 576
30.72\sim 30.72 74 486+384 944
Refer to caption
Figure 10: Ablation analysis on bit allocation schemes between codec-1 and codec-2 in Model-IV at 3232 kbps.

IV-G Complexity and Delay

The proposed NWC model is with 0.350.35 million parameters, a half of which is for the decoder. Hence, in 3232 kbps with two NWC modules for residual coding, the model size totals 0.7M parameters, with the decoder size of 0.350.35M. Even though our decoder is still not as compact as those in traditional codecs, it is 100×100\times smaller than a WaveNet decoder.

Aside from the model size, we investigate the codec's delay and the execution induced latency. The codec will have algorithmic delay if it relies on future samples to predict the current sample. The processing time during the encoding and decoding processes also adds up to the runtime overhead.

IV-G1 Algorithmic delay

The delay of our system is defined by the frame size: the first sample of a frame can be processed only after the entire frame is buffered: 1024/16000=641024/16000=64ms. Causal convolution can minimize such delay at the expense of the reduced speech quality, because it only uses past samples.

TABLE VI: Execution time ratios during model inference (%).
Hardware 0.45M 0.35M 0.45M×\times2 0.35M×\times2
1×1\times Tesla V100 12.49 13.38 20.69 21.12
1×1\times Tesla K80 24.45 22.53 39.42 38.82
8×8\times CPU cores 20.76 18.91 35.17 33.80
1×1\times CPU core 46.88 42.44 87.38 80.21

IV-G2 Analysis of the execution time

The execution time is another important factor to be considered for real-time communications. The bottom line is that the execution of the encoding and decoding processes is expected to be within the duration of the hop length so as not to lead to execution induced latency. For example, WaveNet codec [20] minimizes the system delay using causal convolution, but its processing time, though not reported, can be rather high as it is an autoregressive model with over 2020 million parameters. TABLE VI lists the execution time ratio of our models. The ratio (in percentage) is defined as the execution time to encode and decode the test signals divided by the duration of those signals. Meanwhile, Kankanahalli's model requires 4.784.78ms to encode and decode a hop length of 3030ms on an NVIDIA® GeForce® GTX 1080 Ti GPU, and 21.4221.42ms on an Intel® CoreTM i7-4970K CPU (3.8GHz), which amount to 15.93%15.93\% and 71.40%71.40\% of the execution time ratio, respectively [38]. Our small-sized models (0.45M and 0.35M) on both CPU (Intel® Xeon® Processor E5-2670 V3 2.3GHz) and GPU run faster than Kankanahali's, while the direct comparison is not fair due to the different computing environment. The CMRL models with two NWC modules require more execution time. Note that all our models compared in this test achieved the real-time processing goal as their ratios are under 100%. The proposed NWC with 0.35 million parameters runs faster on CPUs than its predecessor [25][26] with 0.45 million parameters, although the comparison is not consistent on GPUs, due to TensorFlow's optimization effects at runtime.

IV-G3 Implementation notes and limitations

While the proposed model noticeably reduces the decoding time and memory footprint compared to our previous models and the WaveNet decoder, it may not be ready to directly meet real-world requirements. Concretely, our neural codec can be implemented in a more hardware-friendly fashion, so as to allow the processor to handle multiple other tasks. One promising direction is to quantize the neural network. For example, instead of using the single-precision (32 bit) floating points to represent model weights, we can represent each weight by one of 255 values (8-bit quantization), which enables much simpler arithmetic operations during inference. Furthermore, depending on the model architecture, pruning away less important weights is a sensible method to compress a network. Investigating these network compression methods for our codecs is beyond the scope of this paper, but it has been observed that quantization and pruning bring little to no degradation to neural networks for speech recognition [51]. The recently proposed PercepNet architecture also showed high-quality, real-time speech enhancement is possible using less than 5%5\% of a CPU core for speech enhancement [52]. The current system is not causal with an algorithmic delay of 64ms, which necessitates the employment of causal convolutions for real-time applications.

V Concluding Remarks

Recent neural waveform codecs have outperformed the conventional codecs in terms of coding efficiency and speech quality, at the expense of model complexity. We proposed a scalable and lightweight neural acoustic processing unit for waveform coding. Our smallest model contains only 0.350.35 million parameters whose decoder with 0.120.12 million parameters is more than 160×160\times smaller than the WaveNet based codec. Having a compact design as a neural network, by incorporating a trainable LPC analyzer and residual cascading, our models reconstruct clean English speech samples with the quality on par with or superior to that from standardized codecs. Admittedly, these standardized codecs are more computationally efficient and have reasonably well performed from narrow to full band scenarios already. It is still highly desired if these standalone DSP components can be reformulated into a lightweight but end-to-end trainable format for a full neural speech processing pipeline. To that end, the proposed system serves a candidate as it operates in a frame-wise manner with the processing time less than the frame length, even with a less optimized python-based implementations.

The proposed system is still computationally heavier than traditional speech codecs. Therefore, running the model in embedded systems with limited computational resources may require further model compression, such as parameter quantization and pruning. Although the neural waveform codec is generic and not contingent upon language specific priors, its generalizability to different languages and noisy and reverberant acoustic environments is not guaranteed, especially if the model sizes are relatively small.

We open-sourced the project. The source code and sound examples can be found at: https://saige.sice.indiana.edu/research-projects/neural-audio-coding.

References

  • [1] J.-M. Valin and J. Skoglund, ``A real-time wideband neural vocoder at 1.6 kb/s using LPCNet,'' in Proceedings of the Annual Conference of the International Speech Communication Association (Interspeech), 2019.
  • [2] M. R. Schroeder, ``Vocoders: Analysis and synthesis of speech,'' Proceedings of the IEEE, vol. 54, no. 5, pp. 720–734, 1966.
  • [3] A. V. McCree and T. P. Barnwell, ``A mixed excitation LPC vocoder model for low bit rate speech coding,'' IEEE Transactions on Speech and audio Processing, vol. 3, no. 4, pp. 242–250, 1995.
  • [4] M. Dietz, M. Multrus, V. Eksler, V. Malenovsky, E. Norvell, H. Pobloth, L. Miao, Z. Wang, L. Laaksonen, A. Vasilache, et al., ``Overview of the EVS codec architecture,'' in 2015 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).   IEEE, 2015, pp. 5698–5702.
  • [5] B. Bessette, R. Salami, R. Lefebvre, M. Jelinek, J. Rotola-Pukkila, J. Vainio, H. Mikkola, and K. Jarvinen, ``The adaptive multirate wideband speech codec (AMR-WB),'' IEEE transactions on speech and audio processing, vol. 10, no. 8, pp. 620–636, 2002.
  • [6] J. Stachurski and A. McCree, ``Combining parametric and waveform-matching coders for low bit-rate speech coding,'' in 2000 10th European Signal Processing Conference.   IEEE, 2000, pp. 1–4.
  • [7] M. Schroeder and B. Atal, ``Code-excited linear prediction (CELP): High-quality speech at very low bit rates,'' in Acoustics, Speech, and Signal Processing, IEEE International Conference on ICASSP'85., vol. 10.   IEEE, 1985, pp. 937–940.
  • [8] J. Stachurski and A. McCree, ``A 4 kb/s hybrid MELP/CELP coder with alignment phase encoding and zero-phase equalization,'' in ICASSP'20, IEEE International Conference on Acoustics, Speech, and Signal Processing.   IEEE, 2000, pp. 1379–1382.
  • [9] S. Bruhn, H. Pobloth, M. Schnell, B. Grill, J. Gibbs, L. Miao, K. Järvinen, L. Laaksonen, N. Harada, N. Naka, et al., ``Standardization of the new 3GPP EVS codec,'' in 2015 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).   IEEE, 2015, pp. 5703–5707.
  • [10] A. Rämö and H. Toukomaa, ``Subjective quality evaluation of the 3GPP EVS codec,'' in 2015 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).   IEEE, 2015, pp. 5157–5161.
  • [11] J. Skoglund and J.-M. Valin, ``Improving opus low bit rate quality with neural speech synthesis,'' arXiv preprint arXiv:1905.04628, 2019.
  • [12] A. van den Oord, O. Vinyals, and K. Kavukcuoglu, ``Neural discrete representation learning,'' in Advances in Neural Information Processing Systems (NIPS), 2017, pp. 6306–6315.
  • [13] J. Chorowski, R. J. Weiss, S. Bengio, and A. van den Oord, ``Unsupervised speech representation learning using WaveNet autoencoders,'' IEEE/ACM transactions on audio, speech, and language processing, vol. 27, no. 12, pp. 2041–2053, 2019.
  • [14] Y. Xu, J. Du, L.-R. Dai, and C.-H. Lee, ``A regression approach to speech enhancement based on deep neural networks,'' IEEE/ACM Transactions on Audio, Speech, and Language Processing, vol. 23, no. 1, pp. 7–19, 2014.
  • [15] Y. Luo and N. Mesgarani, ``TasNet: Surpassing ideal time-frequency masking for speech separation,'' IEEE/ACM transactions on audio, speech, and language processing, vol. 27, no. 8, pp. 1256–1266, 2019.
  • [16] A. Nagrani, J. S. Chung, and A. Zisserman, ``Voxceleb: a large-scale speaker identification dataset,'' arXiv preprint arXiv:1706.08612, 2017.
  • [17] A. Graves, A.-R. Mohamed, and G. Hinton, ``Speech recognition with deep recurrent neural networks,'' in 2013 IEEE international conference on acoustics, speech and signal processing.   Ieee, 2013, pp. 6645–6649.
  • [18] O. Abdel-Hamid, A.-R. Mohamed, H. Jiang, L. Deng, G. Penn, and D. Yu, ``Convolutional neural networks for speech recognition,'' IEEE/ACM Transactions on audio, speech, and language processing, vol. 22, no. 10, pp. 1533–1545, 2014.
  • [19] A. Van Den Oord, S. Dieleman, H. Zen, K. Simonyan, O. Vinyals, A. Graves, N. Kalchbrenner, A. W. Senior, and K. Kavukcuoglu, ``WaveNet: A generative model for raw audio.'' Speech Synthesis Workshop, vol. 125, 2016.
  • [20] W. B. Kleijn, F. S. C. Lim, A. Luebs, J. Skoglund, F. Stimberg, Q. Wang, and T. C. Walters, ``WaveNet based low rate speech coding,'' in Proceedings of the IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP), 2018, pp. 676–680.
  • [21] D. P. Kingma and M. Welling, ``Auto-encoding variational bayes,'' arXiv preprint arXiv:1312.6114, 2013.
  • [22] Y. L. C. Garbacea, A. van den Oord, ``Low bit-rate speech coding with VQ-VAE and a WaveNet decoder,'' in Proceedings of the IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP), 2019.
  • [23] J.-M. Valin and J. Skoglund, ``LPCNet: Improving neural speech synthesis through linear prediction,'' in IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).   IEEE, 2019, pp. 5891–5895.
  • [24] N. Kalchbrenner, E. Elsen, K. Simonyan, S. Noury, N. Casagrande, E. Lockhart, F. Stimberg, A. Oord, S. Dieleman, and K. Kavukcuoglu, ``Efficient neural audio synthesis,'' in International Conference on Machine Learning.   PMLR, 2018, pp. 2410–2419.
  • [25] K. Zhen, J. Sung, M. S. Lee, S. Beack, and M. Kim, ``Cascaded cross-module residual learning towards lightweight end-to-end speech coding,'' in Proceedings of the Annual Conference of the International Speech Communication Association (Interspeech), 2019.
  • [26] K. Zhen, M. S. Lee, J. Sung, S. Beack, and M. Kim, ``Efficient and scalable neural residual waveform coding with collaborative quantization,'' in ICASSP 2020-2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).   IEEE, 2020, pp. 361–365.
  • [27] C. Un and D. Magill, ``The residual-excited linear prediction vocoder with transmission rate below 9.6 kbits/s,'' IEEE transactions on communications, vol. 23, no. 12, pp. 1466–1474, 1975.
  • [28] K. Zhen, M. S. Lee, J. Sung, S. Beack, and M. Kim, ``Psychoacoustic calibration of loss functions for efficient end-to-end neural audio coding,'' IEEE Signal Processing Letters, pp. 1–1, 2020.
  • [29] K. He, X. Zhang, S. Ren, and J. Sun, ``Deep residual learning for image recognition,'' in Proceedings of the IEEE conference on computer vision and pattern recognition, 2016, pp. 770–778.
  • [30] Y. N. Dauphin, A. Fan, M. Auli, and D. Grangier, ``Language modeling with gated convolutional networks,'' in Proceedings of the International Conference on Machine Learning (ICML), 2017, pp. 933–941.
  • [31] F. Yu and V. Koltun, ``Multi-scale context aggregation by dilated convolutions,'' arXiv preprint arXiv:1511.07122, 2015.
  • [32] K. Tan, J. T. Chen, and D. L. Wang, ``Gated residual networks with dilated convolutions for supervised speech separation,'' in Proc. ICASSP, 2018.
  • [33] ——, ``Gated residual networks with dilated convolutions for monaural speech enhancement,'' IEEE/ACM Transactions on Audio, Speech, and Language Processing, vol. 27, no. 1, pp. 189–198, 2019.
  • [34] F. Chollet, ``Xception: Deep learning with depthwise separable convolutions,'' in Proceedings of the IEEE conference on computer vision and pattern recognition, 2017, pp. 1251–1258.
  • [35] W. Shi, J. Caballero, F. Huszár, J. Totz, A. P. Aitken, R. Bishop, D. Rueckert, and Z. Wang, ``Real-time single image and video super-resolution using an efficient sub-pixel convolutional neural network,'' in Proceedings of the IEEE conference on computer vision and pattern recognition, 2016, pp. 1874–1883.
  • [36] M. Garey, D. Johnson, and H. Witsenhausen, ``The complexity of the generalized Lloyd-max problem (corresp.),'' IEEE Transactions on Information Theory, vol. 28, no. 2, pp. 255–256, 1982.
  • [37] E. Agustsson, F. Mentzer, M. Tschannen, L. Cavigelli, R. Timofte, L. Benini, and L. V. Gool, ``Soft-to-hard vector quantization for end-to-end learning compressible representations,'' in Advances in Neural Information Processing Systems (NIPS), 2017, pp. 1141–1151.
  • [38] S. Kankanahalli, ``End-to-end optimized speech coding with deep neural networks,'' in IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 2018.
  • [39] F. Itakura, ``Early developments of LPC speech coding techniques,'' in icslp, 1990, pp. 1409–1410.
  • [40] I. R. Titze and D. W. Martin, ``Principles of voice production,'' 1998.
  • [41] D. O'Shaughnessy, ``Linear predictive coding,'' IEEE potentials, vol. 7, no. 1, pp. 29–32, 1988.
  • [42] J. Franke, ``A levinson-durbin recursion for autoregressive-moving average processes,'' Biometrika, vol. 72, no. 3, pp. 573–581, 1985.
  • [43] ITU-T G.722.2:, ``Wideband coding of speech at around 16 kbit/s using adaptive multi-rate wideband (AMR-WB),'' 2003.
  • [44] A. Gersho and V. Cuperman, ``Vector quantization: A pattern-matching technique for speech coding,'' IEEE Communications magazine, vol. 21, no. 9, pp. 15–21, 1983.
  • [45] F. Soong and B. Juang, ``Line spectrum pair (LSP) and speech data compression,'' in ICASSP'84. IEEE International Conference on Acoustics, Speech, and Signal Processing, vol. 9.   IEEE, 1984, pp. 37–40.
  • [46] A. Rix, J. Beerends, M. Hollier, and A. Hekstra, ``Perceptual evaluation of speech quality (PESQ)-a new method for speech quality assessment of telephone networks and codecs,'' in Acoustics, Speech, and Signal Processing, 2001. Proceedings.(ICASSP'01). 2001 IEEE International Conference on, vol. 2.   IEEE, 2001, pp. 749–752.
  • [47] ITU-R Recommendation BS 1534-1, ``Method for the subjective assessment of intermediate quality levels of coding systems (MUSHRA),'' 2003.
  • [48] J. S. Garofolo, L. F. Lamel, W. M. Fisher, J. G. Fiscus, D. S. Pallett, N. L. Dahlgren, and V. Zue, ``TIMIT acoustic-phonetic continuous speech corpus,'' Linguistic Data Consortium, Philadelphia, 1993.
  • [49] D. Kingma and J. Ba, ``Adam: A method for stochastic optimization,'' in Proceedings of the International Conference on Learning Representations (ICLR), 2015.
  • [50] J.-M. Valin, G. Maxwell, T. B. Terriberry, and K. Vos, ``High-quality, low-delay music coding in the Opus codec,'' arXiv preprint arXiv:1602.04845, 2016.
  • [51] S. Han, J. Kang, H. Mao, Y. Hu, X. Li, Y. Li, D. Xie, H. Luo, S. Yao, Y. Wang, H. Yang, and W. J. Dally, ``Ese: Efficient speech recognition engine with sparse lstm on fpga,'' in Proceedings of the 2017 ACM/SIGDA International Symposium on Field-Programmable Gate Arrays.   New York, NY, USA: Association for Computing Machinery, 2017, p. 75–84.
  • [52] J.-M. Valin, U. Isik, N. Phansalkar, R. Giri, K. Helwani, and A. Krishnaswamy, ``A perceptually-motivated approach for low-complexity, real-time enhancement of fullband speech,'' in Proceedings of the Annual Conference of the International Speech Communication Association (Interspeech), 2020.