DeepA: a deep neural analyzer for speech and singing vocoding
Abstract
Conventional vocoders are commonly used as analysis tools to provide interpretable features for downstream tasks such as speech synthesis and voice conversion. They are built under certain assumptions about the signals following signal processing principle, therefore, not easily generalizable to different audio, for example, from speech to singing. In this paper, we propose a deep neural analyzer, denoted as DeepA – a neural vocoder that extracts F0 and timbre/aperiodicity encoding from the input speech that emulate those defined in conventional vocoders. Therefore, the resulting parameters are more interpretable than other latent neural representations. At the same time, as the deep neural analyzer is learnable, it is expected to be more accurate for signal reconstruction and manipulation, and generalizable from speech to singing. The proposed neural analyzer is built based on a variational auto-encoder (VAE) architecture. We show that DeepA improves F0 estimation over the conventional vocoder (WORLD). To our best knowledge, this is the first study dedicated to the development of a neural framework for extracting learnable vocoder-like parameters.
Index Terms— neural vocoder, deep analysis, VAE
1 Introduction
Nowadays deep neural networks are used extensively in the fields of speech synthesis [1, 2], voice conversion [3, 4], singing voice synthesis [5, 6] and so on. The neural solutions continue to use conventional vocoders, such as STRAIGHT [7] and WORLD [8], as they allow one to transform speech to and from a set of parameters, which can be manipulated independently. One common example of vocoder usage within a neural TTS framework is to train the network that maps input text into frames of vocoder parameters. For the training phase vocoder provides features that were proven to be an effective speech representation and after the model is trained the vocoder is effectively used as a synthesizer.
We note that conventional vocoders are designed under some assumptions about the properties of speech signals. A vocoder typically performs two functions – signal analysis and reconstruction. Recent studies show that neural reconstruction techniques, such as WaveNet [9] and WaveRNN [10], have surpassed conventional vocoders in terms of speech quality. This can be attributed to the data-driven techniques that allow the reconstruction function to learn from actual data distributions.
The same strategy for neural reconstruction can naturally be applied for neural analysis. While there are some studies in related topics [11, 12, 13, 14], there is no study on a complete neural vocoder, that transforms input speech into a set of independent parameters which can then be used for speech reconstruction in a single neural framework. We have good reasons to believe that data-driven neural vocoder has its advantage over conventional vocoder, which will be the focus of this paper.
While the quality of the analysis part of the vocoder is sufficient for most of the tasks in the field of speech processing, there are tasks in the field that conventional vocoders struggle with. One such task is singing voice synthesis where the reconstructed singing quality dropped significantly when F0 is changed by more than an octave scale. Investigating this problem carefully led us to the following conclusion: vocoder parameters do not allow good quality speech reconstruction for the case of significant change in F0. There turned out to be three main possible causes for that:
-
1.
inconsistency of the vocoder parameters
-
2.
errors in F0 estimation
-
3.
the assumption that timbre is independent from pitch
The third point (cause 3) above is a result of the fact that conventional vocoders were designed as speech encoding and processing tools – in case of low F0 modulation the assumption holds. Further experiments implied that in order to account for the timbre dependency on pitch linguistic information of the processed speech sample is required. Causes 1 and 2, however, arise from the errors in vocoder algorithm. They should be leveraged by learnable features and a sufficient amount of appropriate data, which motivates the studies in this paper. In this paper, we make a step towards a neural analyzer suitable for singing vocoding by addressing points 1 and 2 from the list above.
The main contribution of this paper is the development a neural speech analyzer, i.e., DeepA, that generates consistent and concise features. DeepA allows for a high-quality singing voice reconstruction and outperforms the conventional vocoder in terms of F0 estimation. To our best knowledge, this is the first study dedicated to neural analyzer-synthesizer pipeline.
This paper is organized as follows: section II is dedicated to the related work, section III provides a more detailed analysis of the issues mentioned above, section IV describes the proposed framework, in section V experimental results and their interpretation are presented. Finally section VII concludes the study.
2 Related work
Before an in-depth problem analysis, it is useful to review both conventional and neural solutions to speech analysis and reconstruction to motivate our proposal. Conventional vocoders were originally developed as speech processing tools for voice compression and transformation (hence the name “vocoder”, which stands for “voice encoder”). They represent input signals with vocoder parameters, and reconstruct output signals from parameters. The parameterization allows for manipulation of the physical properties of input signals. The two functions are referred to as analysis and synthesis.
2.1 Conventional vocoders: analysis & synthesis
The early vocoders did not achieve good quality of speech reconstruction, however that changed with the development of STRAIGHT [7] and then WORLD [8]. They achieve this by decomposing speech into three temporal sequences of parameters – the fundamental frequency (F0), smoothed spectrogram (roughly corresponds to the voice timbre) and aperiodicity (correlates with “voice quality” attribute [15]). This strategy allowed for high-quality speech reconstruction and (within a certain range) manipulation.
Conventional vocoders are still used today, mainly either as voice manipulation tools (for example, for changing F0 in voice conversion tasks) or as the provider of interpretable features for training various neural frameworks [16, 17, 18]. However, being speech processing tools built upon certain assumptions about the analyzed signal, conventional vocoders are not flexible enough to be used for high-quality singing voice synthesis. Even though both WORLD and STRAIGHT were either extended to be able to solve the task [15] or applied directly (with some additional tools) [19], both approaches still use a lot of assumptions and heuristics. It should be mentioned that parameter estimation from data via numerical optimization is often employed by those systems, but the number of these parameters is usually quite small, which makes the whole system to rely on the highly constrained model. While a well-engineered model in tandem with a relatively restricted numerical parameter estimation can outperform the more general data-driven neural frameworks in a task the system was designed for, the latter system is more flexible in terms of application and is more adaptable to new data.
In this paper we focus on utilizing these properties of neural frameworks in combination with structural constraints, inspired by conventional vocoders, to learn a compact singing voice representation, which allows for both the accurate speech reconstruction and manipulation.
2.2 Neural vocoders: synthesis
Recent development in the field of neural networks showed that those systems are capable of high-quality speech synthesis, WaveNet [9], WaveGlow [20], WaveRNN [10], GAN-based models [21, 22], LPCNet [23], neural homomorphic vocoder [24] and NSF models [18] being the few examples. Since they reconstruct speech using F0 and/or other acoustic features (e.g. spectrogram, phoneme encoding etc.) a term “neural vocoder” was coined to describe such systems. The architecture and idea behind each neural vocoder may differ (for example, WaveNet generates raw audio, sample by sample, in autoregressive manner, while hn-NSF essentially implements a harmonic-plus-noise model [25] within a neural framework), but all of those systems are flexible in terms of input features.
However, they are not “vocoders” in full sense of the word as they do not perform speech analysis, but rather rely on another module or algorithm, such as STRAIGHT, WORLD, and short-time Fourier transform, for feature extraction. In that regard such frameworks are essentially neural synthesizers, which rely on the assumption that extracted features are adequate and are free from errors. Even though some of the frameworks attempt to mitigate errors in the input features, speech manipulation using such systems remains a challenge. For example, hn-NSF model is able to deal with F0 errors to some extent, but when combining the modified F0 and the original spectrogram as an input, the resulting speech sample has a pitch that is different from what can be inferred from either of the input parameters.
Given this issue and the fact that there appears to be no study dedicated to the development of neural vocoder in full sense, i.e., a complete neural analyzer and synthesizer, the focus of this paper is the development of such a neural vocoder system. In particular, we employ hn-NSF model as a neural synthesizer to be trained jointly with a neural analyzer, architecture of which was inspired by both conventional vocoders and the hn-NSF model itself.
2.3 Variational auto-encoder: representation learning
Auto-encoder [26] is a system that aims to learn the latent, usually – more compact and/or meaningful, representation of the input. Auto-encoders may be viewed as systems that perform dimensionality reduction, but in contrast to principal component analysis the components of the encoding can be constrained or “conditioned” during the training, making the network to converge to the representation with desired properties. For example, the encoding can be conditioned to be “smooth” in a way that makes sense in terms of the task at hand, which may be used for data denoising [27].
Variational auto-encoder (VAE) [28] extends the basic architecture with an additional constraint – the obtained latent representation should resemble a sample from a certain well-known distribution. Gaussian distribution is a common choice for the distribution in question, since an expression for Kullback–Leibler divergence (which is used as part of the loss) between two Gaussian distributions is quite simple and can be derived analytically. The main idea of VAE is to implicitly estimate the distribution of the input data and learn the effective encoding based on that information.
VAE-based models are widely used for learning disentangled representation of some parameters, that may be used for reconstructing the input. For example, it was used for learning the disentangled representation of pitch and timbre of musical instruments [12] and for learning the latent representation of speaking style in a TTS framework [29]. The framework proposed in this paper is built as a variational auto-encoder in order to learn independent representations of harmonic and noise parts of the input sample, which allows one to use the system for speech decomposition, manipulation and reconstruction, similar to the conventional vocoders.

3 Issues with conventional vocoders
We would like to look into two issues with conventional vocoder algorithms, which adversely affect the quality of singing voice manipulation (causes 1 and 2 mentioned in the introduction). The analysis is focused on the WORLD vocoder [8], which is used extensively and shares common characteristics with STRAIGHT [7] [30].
3.1 Inconsistency of vocoder parameters
Vocoder parameters are designed to be independent, but they are expected to be consistent with each other. For example, both F0 and aperiodicity generated by the WORLD vocoder have clearly defined voiced and unvoiced segments and ideally those segments should overlap each other exactly. However, sometimes that is not the case. As shown in Fig. 1, F0 and aperiodicity, calculated by WORLD [8], may predict different “voicedness” (V/UV flag) of the same frame range (highlighted by a grey rectangle in the picture). The sample, reconstructed using these parameters, exhibits distortion in the interval of inconsistency.

3.2 Errors in F0 estimation
Even when F0 is consistent with aperiodicity, the former may still contain noticeable errors. As shown in Fig. 2, the first group of highlighted frames (at the very beginning of the sample) has no periodic structure according to the mel-spectrogram. However, the vocoder estimated F0 of that range as close to the mean pitch of the sample. The second highlighted group exhibits the opposite problem – according to F0 the frames are unvoiced even though the spectrogram clearly shows the continuous periodic structure in that area.

A vocoder seeks to decompose speech or singing vocals into interpretable vocoder parameters. While text-to-speech systems may allow rare mild distortions, the high dynamic range of the singing voice makes those distortions more prominent, which may significantly degrade the perceived singing quality. Such distortions in conventional vocoders usually arise from hard-coded human assumptions in the system. We believe that a neural analyzer with data-driven modeling training could be a solution, that we will discuss next.
4 Deep Neural Analyzer
The proposed neural analyzer, denoted as DeepA for “Deep Analyzer”, is trained jointly with a neural synthesizer. This allows the neural analyzer and synthesizer to interact during training: the synthesizer can guide the analyzer with its speech reconstruction loss, while the analyzer conditions the synthesizer to effectively utilize the latent representation via the modified VAE and spectrogram reconstruction loss.
DeepA follows an encoder-decoder VAE structure, which is coupled with a hn-NSF [18] synthesizer (see Fig. 3) during training. After the system is trained on speech data, the encoder and hn-NSF model can be used as a speech analyzer and synthesizer respectively. The hn-NSF model is chosen as a synthesizer here because it allows for explicit control of F0 when generating speech. With hn-NSF synthesizer, as long as the F0 parameter is consistent with the rest of the input parameters, the F0 of the generated speech is guaranteed to match the F0 control signal. This is especially important for singing, where a fine fluctuation of pitch in the singing voice [31] may greatly affect the overall voice quality.

4.1 Encoder
The encoder consists of two parallel C-blocks (see Fig. 4), which map the input mel-spectrogram to the latent representations of the noise (matrix in a diagram) and harmonic (matrix in a diagram) parts of the input. Both matrices are rows by columns, where is the latent dimension and is the number of temporal frames in the input. The full latent representation is defined as a vertical stack of and . It should also be mentioned that the last row of the matrix (which conveniently corresponds to the last row of the latent representation itself) is conditioned to resemble the normalized ground truth F0. The encoder architecture is inspired by a hn-NSF [18] neural synthesizer, which uses two separate blocks for generating harmonic and noise parts of the speech signal independently and then adds them together to produce the final output. The same idea is used for the decoder architecture as well, which is described in the next subsection.
4.2 Decoder
As can be seen from the diagram (Fig. 3), the decoder is structurally similar to encoder, using TC-blocks (see Fig. 4) instead of C-blocks. The latent speech representation , which is passed to the decoder as an input, is decomposed into matrices and , which are then processed by the decoder separately. This structural constraint is proven to be essential for learning disentangled representations of harmonic and noise parts of the spectrogram. The decoder transforms those representations into masks, which are then multiplied by the generated white noise mel-spectrogram (NS block in Fig. 3) or a mel-spectrogram of a harmonic signal with ground truth F0 as a fundamental (HS block in Fig.3).
4.3 Training details
We are particularly interested in the performance of F0 estimation and rendering. Therefore, we conduct experiments on singing voice samples from the NUS-HLT Speak-Sing (NHSS) database [32], which constitutes approximately 7 hours of audio. All samples were downsampled to 16 kHz, mel-spectrograms of order 80 were used as input features, ground truth F0 was extracted using the WORLD [8] vocoder and the latent dimension was experimentally chosen to be 16. The framework also uses the modified VAE loss in order to improve the system stability. The full loss of the system is given by the following expression:
(1) |
where is the averaged over time Kullback-Leibler divergence between and , which is essentially the distance between two distributions (see [28] for derivation) and is a hn-NSF synthesizer loss defined in [18]. Basically, it is a sum of three STFT losses between the ground truth waveform and the generated one, which represents the speech reconstruction loss in frequency domain. The first quadratic term of represents F0 reconstruction loss, and the third quadratic term of represents the VAE reconstruction loss (mean squared error between the input mel-spectrogram and the reconstructed one). Here is the number of temporal frames in the input, and denote the predicted mean and log-variance of the latent representation ; and denote ground truth F0, mel-spectrogram and singing waveform respectively, while the corresponding circumflexed variables ( and ) denote the generated ones. and are positive weights determined experimentally. The discovered working configuration is .
5 Experiments
We perform both objective and subjective evaluation. The objective evaluation was performed on a subset of 195 utterances from NHSS dataset [32] (not seen by the system during training), reported in terms of mel-cepstral distortion (MCD) and F0 deviation, computed in two different ways (RMSE and median of absolute difference). The subjective evaluation was performed in a form of a mean-opinion score (MOS) test based on the overall voice quality.
5.1 Experimental setup
In this paper three systems are compared: WORLD vocoder, hn-NSF synthesizer using WORLD vocoder parameters, and the proposed neural analyzer followed by hn-NSF synthesizer. WORLD vocoder parameters were extracted using 5 ms frame period and 1024 frequency bins (for spectrogram and aperiodicity), a baseline hn-NSF synthesizer used F0 extracted by WORLD and mel-spectrogram of order 80 and DeepA used just a mel-spectrogram of order 80 with the hn-NSF synthesizer within the proposed framework using latent spectrogram representation of order 32 as its input.
5.2 Objective evaluation
MCD:
Mel-cepstral distortion (MCD) is a perceptual-based measure of similarity between two speech samples [33]. The MCD statistics for the three systems are summarized in Table 1. It is observed that all the tested models achieved comparable quality in terms of of speech reconstruction. As the MCD metric was shown to correlate with perceived speech quality [33], this result gives a rough idea of the actual quality of the proposed framework relative to the established models.
System | Analyzer | Synthesizer | MCD |
---|---|---|---|
1 | WORLD | WORLD | 5.2 ± 1.0 |
2 | WORLD | hn-NSF | 4.2 ± 0.6 |
3 | DeepA | hn-NSF | 5.1 ± 0.5 |

F0 deviation:
Two methods were used to calculate the deviation between F0 extracted by WORLD and F0 generated by the proposed model – root mean square error (RMSE) and median of absolute difference (further referred to as MD). The expression for MD metric is given below:
(2) |
where and are fundamental frequency contours extracted by WORLD and generated by the proposed model respectively. The results of F0 deviation are as follows:
RMSE: | 28.0 ± 13.0 Hz |
---|---|
MD: | 4.0 ± 1.2 Hz |
As can be seen from F0 deviation values, the RMSE value appears to be quite large. This is due to the fact that RMSE is sensitive to outliers. When several samples with the highest RMSE values were investigated, it turned out that the large error is caused by those frames, for which the WORLD vocoder and DeepA predict the opposite V/UV flag. This made RMSE a very useful indicator to highlight the discrepancy between the two systems. By inspecting 20 worst samples in terms of RMSE, it turned out that in almost all the cases the proposed model predicted the correct V/UV flag (two examples of the F0 contour comparison are shown in Fig. 5).
To accurately represent the overall performance of the proposed framework, the median of absolute difference (MD) was employed. MD is insensitive to outliers, thus giving the robust approximation of the central tendency. The MD value for the DeepA is 4 Hz, which, in conjunction with the observations mentioned in the paragraph above, shows that the system is both in good agreement with WORLD vocoder and robust, predicting more accurate results when WORLD fails.
5.3 Subjective evaluation
We report the mean-opinion score (MOS) in Table 2. This test seeks to determine the overall quality of speech generated by WORLD, hn-NSF (which uses WORLD F0 and mel-spectrogram as input) and the proposed framework (DeepA analyzer + hn-NSF synthesizer) with respect to the ground truth. 10 subjects participated in the test and each listened to 96 utterances in total. From Table 2 it can be seen, that the proposed framework achieved the quality comparable to the other systems. These results together with the objective evaluation in the previous subsection suggest that the proposed framework is capable of both analyzing and reconstructing singing voice with appropriate quality. The advantage of the neural analyzer-synthesizer pipeline lies in the fact that it employs interpretable latent representation that allows direct F0 manipulation in the same way as conventional vocoders.
System | Analyzer | Synthesizer | MOS |
---|---|---|---|
Ground truth | 4.1 ± 0.5 | ||
1 | WORLD | WORLD | 3.8 ± 0.5 |
2 | WORLD | hn-NSF | 3.9 ± 0.5 |
3 | DeepA | hn-NSF | 3.7 ± 0.6 |
6 Conclusion
We propose DeepA, a neural analyzer-synthesizer pipeline as a novel vocoder. It was able to achieve singing quality comparable with WORLD and hn-NSF model in both objective and subjective evaluations. We have seen that DeepA provides more accurate and robust F0 estimation than WORLD, that is required for singing vocoding. As the performance of neural analyzer is on par with that of WORLD, this work marks an important step towards an interpretable neural vocoder.
References
- [1] Heiga Zen, Andrew Senior, and Mike Schuster, “Statistical parametric speech synthesis using deep neural networks,” in 2013 IEEE International Conference on Acoustics, Speech and Signal Processing, 2013, pp. 7962–7966.
- [2] Sercan Ömer Arik, Mike Chrzanowski, Adam Coates, Greg Diamos, Andrew Gibiansky, Yongguo Kang, Xian Li, John Miller, Jonathan Raiman, Shubho Sengupta, and Mohammad Shoeybi, “Deep voice: Real-time neural text-to-speech,” CoRR, vol. abs/1702.07825, 2017.
- [3] Berrak Sisman, Junichi Yamagishi, Simon King, and Haizhou Li, “An overview of voice conversion and its challenges: From statistical modeling to deep learning,” IEEE/ACM Transactions on Audio, Speech, and Language Processing, vol. 29, pp. 132–157, 2021.
- [4] Zhengqi Wen, Kehuang Li, Jianhua Tao, and Chin-Hui Lee, “Deep neural network based voice conversion with a large synthesized parallel corpus,” in 2016 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA), 2016, pp. 1–5.
- [5] Masanari Nishimura, Kei Hashimoto, Keiichiro Oura, Y. Nankaku, and K. Tokuda, “Singing voice synthesis based on deep neural networks,” in INTERSPEECH, 2016.
- [6] Yuan-Hao Yi, Yang Ai, Zhen-Hua Ling, and Li-Rong Dai, “Singing voice synthesis using deep autoregressive neural networks for acoustic modeling,” in Interspeech 2019, 20th Annual Conference of the International Speech Communication Association, Graz, Austria, 15-19 September 2019, Gernot Kubin and Zdravko Kacic, Eds. 2019, pp. 2593–2597, ISCA.
- [7] Hideki Kawahara, “STRAIGHT, exploitation of the other aspect of vocoder: Perceptually isomorphic decomposition of speech sounds,” Acoust Sci Technol, vol. 27349, 11 2006.
- [8] Masanori Morise, Fumiya Yokomori, and Kenji Ozawa, “WORLD: A vocoder-based high-quality speech synthesis system for real-time applications,” IEICE Transactions on Information and Systems, vol. E99.D, pp. 1877–1884, 07 2016.
- [9] Aäron van den Oord, Sander Dieleman, Heiga Zen, Karen Simonyan, Oriol Vinyals, Alex Graves, Nal Kalchbrenner, Andrew W. Senior, and Koray Kavukcuoglu, “Wavenet: A generative model for raw audio,” in The 9th ISCA Speech Synthesis Workshop, Sunnyvale, CA, USA, 13-15 September 2016. 2016, p. 125, ISCA.
- [10] Nal Kalchbrenner, Erich Elsen, Karen Simonyan, Seb Noury, Norman Casagrande, Edward Lockhart, Florian Stimberg, Aäron van den Oord, Sander Dieleman, and Koray Kavukcuoglu, “Efficient neural audio synthesis,” CoRR, vol. abs/1802.08435, 2018.
- [11] Jordi Bonada and Merlijn Blaauw, “Hybrid neural-parametric f0 model for singing synthesis,” in ICASSP 2020 - 2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 2020, pp. 7244–7248.
- [12] Yin-Jyun Luo, Kat Agres, and Dorien Herremans, “Learning disentangled representations of timbre and pitch for musical instrument sounds using gaussian mixture variational autoencoders,” CoRR, vol. abs/1906.08152, 2019.
- [13] Wen-Chin Huang, Yi-Chiao Wu, Chen-Chou Lo, Patrick Lumban Tobing, Tomoki Hayashi, Kazuhiro Kobayashi, Tomoki Toda, Yu Tsao, and Hsin-Min Wang, “Investigation of F0 Conditioning and Fully Convolutional Networks in Variational Autoencoder Based Voice Conversion,” in Proc. Interspeech 2019, 2019, pp. 709–713.
- [14] Jesse H. Engel, Lamtharn Hantrakul, Chenjie Gu, and Adam Roberts, “DDSP: differentiable digital signal processing,” CoRR, vol. abs/2001.04643, 2020.
- [15] Hideki Kawahara and Masanori Morise, “Analysis and synthesis of strong vocal expressions: Extension and application of audio texture features to singing voice,” in 2012 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 2012, pp. 5389–5392.
- [16] Chin-Cheng Hsu, Hsin-Te Hwang, Yi-Chiao Wu, Yu Tsao, and Hsin-min Wang, “Voice conversion from non-parallel corpora using variational auto-encoder,” 2016 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA), pp. 1–6, 2016.
- [17] Kazuhiro Kobayashi, Shinnosuke Takamichi, Satoshi Nakamura, and T. Toda, “The nu-naist voice conversion system for the voice conversion challenge 2016,” in INTERSPEECH, 2016.
- [18] Xin Wang, Shinji Takaki, and Junichi Yamagishi, “Neural source-filter waveform models for statistical parametric speech synthesis,” IEEE/ACM Transactions on Audio, Speech, and Language Processing, vol. 28, pp. 402–415, 2020.
- [19] Ameya/Ayame, “[Vocal Synthesis Tool UTAU] (in Japanese),” http://utau2008.web.fc2.com/index.html, 2008, Retreived: 2010-07-09.
- [20] Ryan Prenger, Rafael Valle, and Bryan Catanzaro, “Waveglow: A flow-based generative network for speech synthesis,” CoRR, vol. abs/1811.00002, 2018.
- [21] Kundan Kumar, Rithesh Kumar, T. Boissière, L. Gestin, Wei Zhen Teoh, Jose M. R. Sotelo, A. D. Brébisson, Yoshua Bengio, and Aaron C. Courville, “Melgan: Generative adversarial networks for conditional waveform synthesis,” in NeurIPS, 2019.
- [22] Jinhyeok Yang, Junmo Lee, Youngik Kim, Hoon-Young Cho, and Injung Kim, “VocGAN: A High-Fidelity Real-Time Vocoder with a Hierarchically-Nested Adversarial Network,” in Proc. Interspeech 2020, 2020, pp. 200–204.
- [23] J. Valin and J. Skoglund, “LPCNet: Improving Neural Speech Synthesis through Linear Prediction,” ICASSP 2019 - 2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 5891–5895, 2019.
- [24] Zhijun Liu, Kuan Chen, and Kai Yu, “Neural Homomorphic Vocoder,” in Proc. Interspeech 2020, 2020, pp. 240–244.
- [25] Y. Stylianou, J. Laroche, and É. Moulines, “High-quality speech modification based on a harmonic + noise model,” in EUROSPEECH, 1995.
- [26] M. Mougeot, R. Azencott, and B. Angeniol, “Image compression with back propagation: Improvement of the visual restoration using different cost functions,” Neural Networks, vol. 4, no. 4, pp. 467–476, 1991.
- [27] Xiao-Jiao Mao, Chunhua Shen, and Yu-Bin Yang, “Image restoration using convolutional auto-encoders with symmetric skip connections,” CoRR, vol. abs/1606.08921, 2016.
- [28] Diederik P. Kingma and M. Welling, “Auto-encoding variational bayes,” CoRR, vol. abs/1312.6114, 2014.
- [29] Ya-Jie Zhang, Shifeng Pan, Lei He, and Zhen-Hua Ling, “Learning latent representations for style control and transfer in end-to-end speech synthesis,” in ICASSP 2019 - 2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 2019, pp. 6945–6949.
- [30] Hideki Kawahara, Masanori Morise, and Kanru Hua, “Revisiting spectral envelope recovery from speech sounds generated by periodic excitation,” in 2018 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA ASC), 2018, pp. 1674–1683.
- [31] Hiroki Mori, W. Odagiri, and Hideki Kasuya, “F0 dynamics in singing: Evidence from the data of a baritone singer,” IEICE Transactions on Information and Systems, pp. 1086–1092, 05 2004.
- [32] Bidisha Sharma, Xiaoxue Gao, Karthika Vijayan, Xiaohai Tian, and Haizhou Li, “NHSS: A speech and singing parallel database,” CoRR, vol. abs/2012.00337, 2020.
- [33] R. Kubichek, “Mel-cepstral distance measure for objective speech quality assessment,” in Proceedings of IEEE Pacific Rim Conference on Communications Computers and Signal Processing, 1993, vol. 1, pp. 125–128 vol.1.