This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Bailing-TTS: Chinese Dialectal Speech Synthesis Towards Human-like Spontaneous Representation

Xinhan Di, Zihao Chen, Yunming Liang, Junjie Zheng, Yihua Wang, Chaofan Ding
AI Lab, Giant Network
Abstract

Large-scale text-to-speech (TTS) models have made significant progress recently. However, they still fall short in the generation of Chinese dialectal speech. To address this, we propose Bailing-TTS, a family of large-scale TTS models capable of generating high-quality Chinese dialectal speech. Bailing-TTS serves as a foundation model for Chinese dialectal speech generation. First, continual semi-supervised learning is proposed to facilitate the alignment of text tokens and speech tokens. Second, the Chinese dialectal representation learning is developed using a specific transformer architecture and multi-stage training processes. With the proposed design of novel network architecture and corresponding strategy, Bailing-TTS is able to generate Chinese dialectal speech from text effectively and efficiently. Experiments demonstrate that Bailing-TTS generates Chinese dialectal speech towards human-like spontaneous representation. Readers are encouraged to listen to demos at https://giantailab.github.io/bailingtts_tech_report/index.html.

1 Introduction

The goal of text-to-speech (TTS) systems [30, 10, 6, 38, 33, 36] is to generate human-like speech from text. With the development of neural network and deep learning, large-scale training corpus of a human-like level [34, 28, 27, 31, 29] is established and corresponding TTS models is developed. However, such system [4, 19, 23, 5, 12] is only able to generate voice for non-dialectal speech. And the quality of the generated speech is not satisfactory [39, 32, 20, 16, 18, 35]. Therefore, in order to generate human-like quality of speech from dialectal text, we propose Bailing-TTS, a family of speech generation models for synthesizing Chinese dialectal speech from text with human-level naturalness.

Besides, in zero-shot speech synthesis, both the naturalness and the robustness of speaker modeling are challenging [26]. In previous studies, large-scale models are developed for speech synthesis[19, 2, 16, 3, 37]. Despite the advances from these methods, the generalization capability of these methods is still insufficient for the generation of high-quality dialectal speech. Therefore, we primarily propose Bailing-TTS to generate high quality Chinese dialectal speech from text.

We propose several specific extension techniques that significantly enhance the quality of Chinese dialectal speech generation. First, continual semi-supervised learning framework towards text and speech tokens is developed to facilitate the alignment of these multiple modalities. Second, a multiple-stage Chinese dialectal representation learning framework introduces a specific network architecture and the corresponding learning strategy to improve the quality of the generated Chinese dialectal speech.

In the evaluation of the proposed Bailing-TTS, both the advantages and disadvantages of the proposed model for the dialectal speech generation is compared with human speakers. The evaluation results represent that Bailing-TTS generates good quality of Chinese dialectal speech.

Finally, both the potential applications and the limitation of Bailing-TTS are discussed. Then, the future work of Bailing-family foundation model is represented including taking multiple modalities as input and producing audio output. For example, it’s supposed to produce the generation of music [7] from the text and video. Besides the generation of audio such as music or speech, in our next step, the video and audio are generated simultaneously.

Our key contributions are represented as the following:

  • We introduce Bailing-TTS, a family of foundation models for synthesizing Chinese dialectal speech from text with human-level naturalness, such as spontaneous speech.

  • Both a continual semi-supervised representation strategy and Chinese dialectal representation with a specific mixture-of-expert network architecture are proposed to improve the quality of the generated Chinese dialectal speech.

  • A variety of hierarchical reinforcement post-learning extension techniques are represented to significantly enhance the quality of Chinese dialectal speech generation.

2 Bailing-TTS

We propose Bailing-TTS based on an auto-regressive transformer model of multiple layers. The proposed Bailing-TTS is trained on a large-scale of dataset including a large part of dialectal data. Both non-precise annotation data and high-quality annotation data are in the dataset. Besides, we propose a multi-stage training strategy, based on a specific transformer architecture (Figure 1), to facilitate the generation of spontaneous and expressive speech from text.

Refer to caption
Figure 1: Overall Architecture of Bailing-TTS

2.1 Overall Architecture

First, at the stage of token representation, a continual semi-supervised learning strategy of spontaneous, expressive text and speech token pairs is proposed to facilitate weak alignment between the two modalities. Second, a specific transformer-based architecture is proposed corresponding with a multi-stage training strategy. Third, during inference, it generates good quality of spontaneous Chinese dialectal speech from text.

2.2 Continual Semi-supervised Learning Towards Text and Speech Tokens

Spontaneous speaking style exhibits notable differences from other speaking styles due to various spontaneous phenomena (e.g., filled pauses, prolongation) and substantial prosody variation (e.g., diverse pitch and duration variation), posing challenges to modeling and prediction of spontaneous style.

In order to generate spontaneous, expressive speech from text, we present a multi-stage and multi-modal (text and speech) pre-training learning framework for text-speech token alignment (Figure 1). In the first stage, an unsupervised sampling strategy is proposed on a large-scale non-precise annotation datasets, forming a multi-stage pre-training process. In the second stage, a fine-tune sampling strategy is proposed on a high-quality dataset containing spontaneous, expressive text and speech pairs. Therefore, the token-wise text-speech association is represented to facilitate alignment between the two modalities.

Refer to caption
Figure 2: Overall Architecture of Dialectal MoE

2.3 Multi-Chinese Dialectal Representation Learning

Dialect-specific TTS models are not satisfactory as dialect-specific data is scarce. Therefore, a single unified TTS model that generalizes well for many dialects is in demand.

In order to train a unified TTS model for a variety of Chinese dialects, we present a specific token representation together with a multi-stage learning strategy. In the aspect of the token representation, a design of mixture-of-expert architecture (Figure 2) is proposed to learn the unified representation of multiple Chinese dialects and specific representation of each dialect. Similarly, we propose a cross-attention based mechanism to inject dialect tokens throughout the TTS model layers.

2.4 Hierarchical RL-based Posting-training Extension

Similar to text-based language models, four training stages are built for the representation learning of Bailing-TTS. The first stage is the pre-training stage aiming to maximize scenario and speaker coverage with a robust backbone for general speech representation of the Chinese dialect. This stage is to learn the efficient representation of a large-scale Chinese dialect dataset. The second stage is the fine-tuning stage in order to enhance the basic performance of the dialectal speech.

Besides, we propose a hierarchical RL learning strategy as the third and the fourth step in order to generate the high-quality speech of multiple Chinese dialects. In the RL hierarchical training strategy, in order to let the high-level policy support exploration for good quality of spontaneous, expressive speech and the primary-level policy support good quality generation of dialectal speech simultaneously. We propose dynamic sub-optimization for the learning of the high-level policy at the base of the primary-level policy in the hierarchical learning strategy.

3 Experiments and Results

3.1 Experimental Settings

In this section, the details of the training, inference and evaluation for the proposed Bailing-TTS on Chinese dialectal speech synthesis are introduced.

3.1.1 Implementation Details

A custom dataset containing high-quality Mandarin and multiple Chinese dialects data is used for the training, inference and evaluation. This dataset contains 200k200k hours of labeled speech data. The annotation contains fine-grained labels between text and speech modalities. In addition, this large-scale dataset contains high-quality speech-text pairs. Furthermore, this custom Chinese dialect dataset is used for the pre-training stage of the continual semi-supervised learning. Clusters of GPUs (A100) are applied for the training.

3.1.2 Evaluation Dataset

A custom test set is used for the Chinese dialect TTS task. There are 5050 distinct Chinese speakers for each Chinese dialect with 20002000 clips. Following [29], sentences are randomly chosen from each speaker, and 5-second to 15-second clips are randomly chosen as prompts from the same speaker’s speech [29]. Besides, there are 2525 male and 2525 female Chinese speakers for each Chinese dialect. In the dataset, each Chinese dialect contains 5050 common spontaneous words [22]. Besides, 20002000 clips from DiDiSpeech dataset [14] are used for the evaluation of Mandarin speech.

3.1.3 Evaluation Metrics

Both objective metrics and subjective metrics are used for the evaluation. In the aspect of objective metrics: word error rate (WER) and speaker similarity (SIM) is evaluated. In the aspect of subjective metrics, comparative mean option scores (CMOS) and mean option score (MOS) are calculated to evaluate the naturalness and quality, respectively.

3.2 Experimental Results on TTS of Multiple Chinese Dialects

In this section, the proposed Bailing-TTS is evaluated in terms of 1)generation quality 2)generation similarity 3)human-level naturalness.

3.2.1 Generation Quality

As shown in Table 1, we find that the proposed Bailing-TTS is close to the ground-truth recording. It demonstrates Bailing-TTS can generate high-quality and natural speech for the testing Chinese dialect data. Besides, the performance verifies the effectiveness of the proposed Bailing-TTS with specific network architecture and the corresponding training strategy. In details, for the evaluation of Mandarin speech, Bailing-TTS obtains 1.861.86 for word error rate (WER), 4.324.32 for mean option score (MOS). In addition, human speakers obtain 1.351.35 for word error rate (WER), 4.324.32 for mean option score (MOS). For the evaluation of dialectal speech, Bailing-TTS obtains 6.376.37 for word error rate (WER), 4.114.11 for mean option score (MOS). Besides, human speakers obtain 4.604.60 for word error rate (WER), 4.244.24 for mean option score (MOS).

Table 1: The evaluation results for Bailing-TTS on Chinese speech dataset.
Systems WER\textbf{WER}\downarrow MOS\textbf{MOS}\uparrow CMOS\textbf{CMOS}\uparrow WER\textbf{WER}\downarrow MOS\textbf{MOS}\uparrow CMOS\textbf{CMOS}\uparrow
Mandarin Dialect
Human 1.351.35 4.324.32 - 4.604.60 4.244.24 -
Bailing-TTS 1.861.86 4.214.21 0.06-0.06 6.376.37 4.114.11 0.08-0.08

3.2.2 Human-Level Naturalness

The speech synthesized through Bailing-TTS is also in the comparison with the Ground Truth in Table 1 for the evaluation of both Mandarin speech and Chinese dialectal speech. Bailing-TTS achieves 0.06-0.06 CMOS and 0.08-0.08 CMOS compared to the Ground Truth, which demonstrates that our method is comparable on human voice quality.

3.2.3 Zero-shot learning

Both objective and subjective evaluation for the zero-shot learning experiments are conducted. In one aspect, clips from custom corpora are used to measure the proposed model’s performance on a variety of objective metrics. 20002000 clips from custom dataset are used for the evaluation of Mandarin speech. To be noted, the custom clips contains highly expressive Mandarin speech with spontaneous word.

Besides, both reference utterance and corresponding target utterance are collected from the same speaker for each clip. The proposed Bailing-TTS family is designed to generate target text based on a prompt (the reference speech). Therefore, comparison between synthesized speech and ground truth speech from real human is conducted for both the objective and subjective evaluation.

For the comparison with real human speech, the word error rate (WER) and speaker similarity metrics (SIM) are used for objective evaluation. Particularly, Paraformer-zh [13] is used as the automatic speech recognition (ASR) for Mandarin, and we train a custom automatic speech recognition (ASR) for multiple Chinese dialects. Besides, comparative mean option scores (CMOS) are used for subjective evaluation. Particularly, first, native Chinese dialectal speakers are represented with a reference speech clip of the target speaker. Second, both the synthesized output of the proposed model and the corresponding targeting human speech are used. Third, evaluation is conducted to rate each given clip with higher speaker similarity and expressiveness to the reference clip on a scale between 3-3 to +3+3, where 3-3 and 33 indicate the least and strongest preference. The results are then collected and averaged over all human evaluators and test sentences.

The results for both objective and subjective evaluation are reported in Table 2. The results demonstrate that the proposed Bailing-TTS model represents an expected performance in zero-shot learning. In details, for the evaluation of Mandarin speech (Zero-shot), Bailing-TTS obtains 3.483.48 for word error rate (WER), 0.720.72 for speaker similarity metrics (SIM).

3.2.4 Fine-tuning learning

Speaker fine-tuning (SFT) is conducted on the base of pre-trained Bailing-TTS model. For the evaluation of the fine-tuning learning, speech data from 1010 speakers are combined for Mandarin fine-tuning learning dataset. Then, the comparison is made between the fine-tuned model (Bailing-TTS fine-tuned) and the base pre-trained model (Bailing-TTS Zero-shot). For the objective evaluation, the word error rate (WER) and speaker similarity metrics (SIM) are used for the objective evaluation. For the subjective evaluation, comparative mean option scores (CMOS) are used. To be noted, 1515 seconds of speech clip is randomly sampled as the prompt for each speaker. Similarly, speech data from 1010 speakers are combined for Chinese dialectal fine-tuning learning dataset.

For the evaluation of Mandarin speech, the results of the speaker fine-tuning experiment and the base model are demonstrated in the Table 2. The results demonstrate that the fine-tuned models shows better performance in both objective and subjective metrics. In details, for the evaluation of Mandarin speech (Speaker fine-tuned), Bailing-TTS obtains 2.982.98 for word error rate (WER), 0.770.77 for speaker similarity metrics (SIM) and +0.18+0.18 for comparative mean option scores (CMOS). For the evaluation of Chinese dialectal speech (Speaker fine-tuned), Bailing-TTS obtains 7.437.43 for word error rate (WER), 0.760.76 for speaker similarity metrics (SIM) and +0.11+0.11 for comparative mean option scores (CMOS).

Table 2: The evaluation results of zero-shot learning and speaker fine-tuning for Bailing-TTS on Chinese speech dataset.
Systems WER\textbf{WER}\downarrow SIM\textbf{SIM}\uparrow CMOS\textbf{CMOS}\uparrow
Mandarin
Bailing-TTS(Zero-shot) 3.483.48 0.720.72 -
Bailing-TTS(Speaker fine-tuned) 2.982.98 0.770.77 +0.18+0.18
Systems Dialect
Bailing-TTS(Speaker fine-tuned) 7.437.43 0.760.76 +0.11+0.11

3.2.5 Streaming processing

There are a range of challenges in the application of TTS systems in the real world. User experiences are not satisfactory due to the latency and the first packet delay. The computation cost of both the time and the memory is still large for the deployment on mobile hardware systems. Therefore, initial work on the deployment is conducted. A variety of techniques [9, 1, 24, 21] are employed to reduce the inference cost and latency.

In order to reduce the inference cost and the memory cost of attention layers, first, we apply efficient memory cost methods including grouped-query attention [1], paged attention [17], and flash attention [9, 8]. Second, model quantization method is also applied to further reduce the computation cost [25, 15]. Similarly, the consistency distillation [30] and a modified flow matching algorithm [11] are used to reduce the computation cost of the diffusion model. Third, to further reduce the computation cost of diffusion architecture, in one aspect, the consistency distillation [30] is applied to reduce the computation cost of the diffusion model, in the other aspect, the flow matching algorithm [11] is used to reduce the memory cost.

The initial results demonstrate that comparable performance is achieved after a variety of methods are applied on the deployed model. In details, as shown in Table 3, the latency is reduced to 0.13×0.13\times, the RTF is reduced to 0.46×0.46\times and WER and SIM remains the same in the comparison with the offline Bailing-TTS model.

Table 3: Comparison between the offline Bailing-TTS model and the online Bailing-TTS model.
System Latency\textbf{Latency}\downarrow RTF\textbf{RTF}\downarrow WER\textbf{WER}\downarrow SIM\textbf{SIM}\uparrow CMOS\textbf{CMOS}\uparrow
Offline Bailing-TTS 1×1\times 1×1\times 2.752.75 0.780.78 -
Online Bailing-TTS 0.13×0.13\times 0.46×0.46\times 2.752.75 0.780.78 0.04-0.04

4 Applications and Discussion

We propose a family of Bailing-TTS towards Chinese dialectal speech synthesis with spontaneous words. It has the potential for real-world application. First, the dialectal speech is promising to provide rich experience of chat service in the real-world particularly companion chat service. Second, it’s very likely to be beneficial to the facilitation of dialectal culture and cultural applications.

Initial work on the family of Bailing-TTS is proposed while services such as speech synthesis with emotions, speech with support of other modalities are not well explored. Therefore, the exploration of expressive and emotional generation of Chinese dialectal speech is under-going. We are planing to develop the next version of Bailing-TTS family for the generation of high-quality audio (speech/music) from the input of video and text. Then, the generation of both the high-quality audio together with video will be explored.

References

  • [1] Joshua Ainslie, James Lee-Thorp, Michiel de Jong, Yury Zemlyanskiy, Federico Lebrón, and Sumit Sanghai. Gqa: Training generalized multi-query transformer models from multi-head checkpoints. arXiv preprint arXiv:2305.13245, 2023.
  • [2] Zalán Borsos, Raphaël Marinier, Damien Vincent, Eugene Kharitonov, Olivier Pietquin, Matt Sharifi, Dominik Roblek, Olivier Teboul, David Grangier, Marco Tagliasacchi, et al. Audiolm: a language modeling approach to audio generation. IEEE/ACM transactions on audio, speech, and language processing, 31:2523–2533, 2023.
  • [3] Zalán Borsos, Matt Sharifi, Damien Vincent, Eugene Kharitonov, Neil Zeghidour, and Marco Tagliasacchi. Soundstorm: Efficient parallel audio generation. arXiv preprint arXiv:2305.09636, 2023.
  • [4] Sanyuan Chen, Chengyi Wang, Zhengyang Chen, Yu Wu, Shujie Liu, Zhuo Chen, Jinyu Li, Naoyuki Kanda, Takuya Yoshioka, Xiong Xiao, et al. Wavlm: Large-scale self-supervised pre-training for full stack speech processing. IEEE Journal of Selected Topics in Signal Processing, 16(6):1505–1518, 2022.
  • [5] Zehua Chen, Yihan Wu, Yichong Leng, Jiawei Chen, Haohe Liu, Xu Tan, Yang Cui, Ke Wang, Lei He, Sheng Zhao, et al. Resgrad: Residual denoising diffusion probabilistic models for text to speech. arXiv preprint arXiv:2212.14518, 2022.
  • [6] Ha-Yeong Choi, Sang-Hoon Lee, and Seong-Whan Lee. Dddm-vc: Decoupled denoising diffusion models with disentangled representation and prior mixup for verified robust voice conversion. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 38, pages 17862–17870, 2024.
  • [7] Geoffrey Cideron, Sertan Girgin, Mauro Verzetti, Damien Vincent, Matej Kastelic, Zalán Borsos, Brian McWilliams, Victor Ungureanu, Olivier Bachem, Olivier Pietquin, et al. Musicrl: Aligning music generation to human preferences. arXiv preprint arXiv:2402.04229, 2024.
  • [8] Tri Dao. Flashattention-2: Faster attention with better parallelism and work partitioning. arXiv preprint arXiv:2307.08691, 2023.
  • [9] Tri Dao, Dan Fu, Stefano Ermon, Atri Rudra, and Christopher Ré. Flashattention: Fast and memory-efficient exact attention with io-awareness. Advances in Neural Information Processing Systems, 35:16344–16359, 2022.
  • [10] Yan Deng, Long Zhou, Yuanhao Yi, Shujie Liu, and Lei He. Prosody-aware speecht5 for expressive neural tts. In ICASSP 2023-2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pages 1–5. IEEE, 2023.
  • [11] Patrick Esser, Sumith Kulal, Andreas Blattmann, Rahim Entezari, Jonas Müller, Harry Saini, Yam Levi, Dominik Lorenz, Axel Sauer, Frederic Boesel, et al. Scaling rectified flow transformers for high-resolution image synthesis. In Forty-first International Conference on Machine Learning, 2024.
  • [12] Yuan Gao, Nobuyuki Morioka, Yu Zhang, and Nanxin Chen. E3 tts: Easy end-to-end diffusion-based text to speech. In 2023 IEEE Automatic Speech Recognition and Understanding Workshop (ASRU), pages 1–8. IEEE, 2023.
  • [13] Zhifu Gao, Zerui Li, Jiaming Wang, Haoneng Luo, Xian Shi, Mengzhe Chen, Yabin Li, Lingyun Zuo, Zhihao Du, Zhangyu Xiao, et al. Funasr: A fundamental end-to-end speech recognition toolkit. arXiv preprint arXiv:2305.11013, 2023.
  • [14] Tingwei Guo, Cheng Wen, Dongwei Jiang, Ne Luo, Ruixiong Zhang, Shuaijiang Zhao, Wubo Li, Cheng Gong, Wei Zou, Kun Han, et al. Didispeech: A large scale mandarin speech corpus. In ICASSP 2021-2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pages 6968–6972. IEEE, 2021.
  • [15] Yi Guo, Fanliu Kong, Xiaoyang Li, Hui Li, Wei Chen, Xiaogang Tian, Jinping Cai, Yang Zhang, and Shouda Liu. decoupleq: Towards 2-bit post-training uniform quantization via decoupling parameters into integer and floating points. arXiv preprint arXiv:2404.12759, 2024.
  • [16] Ziyue Jiang, Yi Ren, Zhenhui Ye, Jinglin Liu, Chen Zhang, Qian Yang, Shengpeng Ji, Rongjie Huang, Chunfeng Wang, Xiang Yin, et al. Mega-tts: Zero-shot text-to-speech at scale with intrinsic inductive bias. arXiv preprint arXiv:2306.03509, 2023.
  • [17] Woosuk Kwon, Zhuohan Li, Siyuan Zhuang, Ying Sheng, Lianmin Zheng, Cody Hao Yu, Joseph Gonzalez, Hao Zhang, and Ion Stoica. Efficient memory management for large language model serving with pagedattention. In Proceedings of the 29th Symposium on Operating Systems Principles, pages 611–626, 2023.
  • [18] Mateusz Łajszczak, Guillermo Cámbara, Yang Li, Fatih Beyhan, Arent van Korlaar, Fan Yang, Arnaud Joly, Álvaro Martín-Cortinas, Ammar Abbas, Adam Michalski, et al. Base tts: Lessons from building a billion-parameter text-to-speech model on 100k hours of data. arXiv preprint arXiv:2402.08093, 2024.
  • [19] Matthew Le, Apoorv Vyas, Bowen Shi, Brian Karrer, Leda Sari, Rashel Moritz, Mary Williamson, Vimal Manohar, Yossi Adi, Jay Mahadeokar, et al. Voicebox: Text-guided multilingual universal speech generation at scale. Advances in neural information processing systems, 36, 2024.
  • [20] Yuxi Li. Deep reinforcement learning: An overview. arXiv preprint arXiv:1701.07274, 2017.
  • [21] Ji Lin, Jiaming Tang, Haotian Tang, Shang Yang, Xingyu Dang, and Song Han. Awq: activationaware weight quantization for llm compression and acceleration. corr, abs/2306.00978, 2023. doi: 10.48550. arXiv preprint ARXIV.2306.00978.
  • [22] Steven R Livingstone and Frank A Russo. The ryerson audio-visual database of emotional speech and song (ravdess): A dynamic, multimodal set of facial and vocal expressions in north american english. PloS one, 13(5):e0196391, 2018.
  • [23] Justin Lovelace, Soham Ray, Kwangyoun Kim, Kilian Q Weinberger, and Felix Wu. Simple-tts: End-to-end text-to-speech synthesis with latent diffusion. 2023.
  • [24] Simian Luo, Yiqin Tan, Longbo Huang, Jian Li, and Hang Zhao. Latent consistency models: Synthesizing high-resolution images with few-step inference. arXiv preprint arXiv:2310.04378, 2023.
  • [25] Markus Nagel, Marios Fournarakis, Rana Ali Amjad, Yelysei Bondarenko, Mart Van Baalen, and Tijmen Blankevoort. A white paper on neural network quantization. arXiv preprint arXiv:2106.08295, 2021.
  • [26] Vadim Popov, Ivan Vovk, Vladimir Gogoryan, Tasnima Sadekova, and Mikhail Kudinov. Grad-tts: A diffusion probabilistic model for text-to-speech. In International Conference on Machine Learning, pages 8599–8608. PMLR, 2021.
  • [27] Yi Ren, Yangjun Ruan, Xu Tan, Tao Qin, Sheng Zhao, Zhou Zhao, and Tie-Yan Liu. Fastspeech: Fast, robust and controllable text to speech. Advances in neural information processing systems, 32, 2019.
  • [28] Jonathan Shen, Ruoming Pang, Ron J Weiss, Mike Schuster, Navdeep Jaitly, Zongheng Yang, Zhifeng Chen, Yu Zhang, Yuxuan Wang, Rj Skerrv-Ryan, et al. Natural tts synthesis by conditioning wavenet on mel spectrogram predictions. In 2018 IEEE international conference on acoustics, speech and signal processing (ICASSP), pages 4779–4783. IEEE, 2018.
  • [29] Kai Shen, Zeqian Ju, Xu Tan, Yanqing Liu, Yichong Leng, Lei He, Tao Qin, Sheng Zhao, and Jiang Bian. Naturalspeech 2: Latent diffusion models are natural and zero-shot speech and singing synthesizers. arXiv preprint arXiv:2304.09116, 2023.
  • [30] Yakun Song, Zhuo Chen, Xiaofei Wang, Ziyang Ma, and Xie Chen. Ella-v: Stable neural codec language modeling with alignment-guided sequence reordering. arXiv preprint arXiv:2401.07333, 2024.
  • [31] Xu Tan, Jiawei Chen, Haohe Liu, Jian Cong, Chen Zhang, Yanqing Liu, Xi Wang, Yichong Leng, Yuanhao Yi, Lei He, et al. Naturalspeech: End-to-end text-to-speech synthesis with human-level quality. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2024.
  • [32] Chengyi Wang, Sanyuan Chen, Yu Wu, Ziqiang Zhang, Long Zhou, Shujie Liu, Zhuo Chen, Yanqing Liu, Huaming Wang, Jinyu Li, et al. Neural codec language models are zero-shot text to speech synthesizers. arXiv preprint arXiv:2301.02111, 2023.
  • [33] Xiaofei Wang, Manthan Thakker, Zhuo Chen, Naoyuki Kanda, Sefik Emre Eskimez, Sanyuan Chen, Min Tang, Shujie Liu, Jinyu Li, and Takuya Yoshioka. Speechx: Neural codec language model as a versatile speech transformer. IEEE/ACM Transactions on Audio, Speech, and Language Processing, 2024.
  • [34] Yuxuan Wang, RJ Skerry-Ryan, Daisy Stanton, Yonghui Wu, Ron J Weiss, Navdeep Jaitly, Zongheng Yang, Ying Xiao, Zhifeng Chen, Samy Bengio, et al. Tacotron: Towards end-to-end speech synthesis. arXiv preprint arXiv:1703.10135, 2017.
  • [35] Zhichao Wang, Yuanzhe Chen, Lei Xie, Qiao Tian, and Yuping Wang. Lm-vc: Zero-shot voice conversion via speech generation based on language models. IEEE Signal Processing Letters, 2023.
  • [36] Lijun Yu, José Lezama, Nitesh B Gundavarapu, Luca Versari, Kihyuk Sohn, David Minnen, Yong Cheng, Agrim Gupta, Xiuye Gu, Alexander G Hauptmann, et al. Language model beats diffusion–tokenizer is key to visual generation. arXiv preprint arXiv:2310.05737, 2023.
  • [37] Neil Zeghidour, Alejandro Luebs, Ahmed Omran, Jan Skoglund, and Marco Tagliasacchi. Soundstream: An end-to-end neural audio codec. IEEE/ACM Transactions on Audio, Speech, and Language Processing, 30:495–507, 2021.
  • [38] Dong Zhang, Zhaowei Li, Shimin Li, Xin Zhang, Pengyu Wang, Yaqian Zhou, and Xipeng Qiu. Speechalign: Aligning speech generation to human preferences. arXiv preprint arXiv:2404.05600, 2024.
  • [39] Ziqiang Zhang, Long Zhou, Chengyi Wang, Sanyuan Chen, Yu Wu, Shujie Liu, Zhuo Chen, Yanqing Liu, Huaming Wang, Jinyu Li, et al. Speak foreign languages with your own voice: Cross-lingual neural codec language modeling. arXiv preprint arXiv:2303.03926, 2023.