MOSPC: MOS Prediction Based on Pairwise Comparison
Abstract
As a subjective metric to evaluate the quality of synthesized speech, Mean opinion score (MOS) usually requires multiple annotators to score the same speech. Such an annotation approach requires a lot of manpower and is also time-consuming. MOS prediction model for automatic evaluation can significantly reduce labor cost. In previous works, it is difficult to accurately rank the quality of speech when the MOS scores are close. However, in practical applications, it is more important to correctly rank the quality of synthesis systems or sentences than simply predicting MOS scores. Meanwhile, as each annotator scores multiple audios during annotation, the score is probably a relative value based on the first or the first few speech scores given by the annotator. Motivated by the above two points, we propose a general framework for MOS prediction based on pair comparison (MOSPC), and we utilize C-Mixup algorithm to enhance the generalization performance of MOSPC. The experiments on BVCC and VCC2018 show that our framework outperforms the baselines on most of the correlation coefficient metrics, especially on the metric KTAU related to quality ranking. And our framework also surpasses the strong baseline in ranking accuracy on each fine-grained segment. These results indicate that our framework contributes to improving the ranking accuracy of speech quality.
1 Introduction
Speech quality evaluation metrics are designed to reflect the speech quality of synthesized speech. Speech quality evaluation metrics include objective metrics Kubichek (1993); Kim (2005); Malfait et al. (2006) and subjective metrics Wester et al. (2016a).
MOS prediction is the task of constructing an automatic evaluation metric by fitting the subjective evaluation metric MOS. The training process of previous works mainly focus on predicting the MOS of a single speech. By reviewing the annotation process of MOS, we found that comparison may be a potential scoring strategy employed by some of the annotators. Specifically, in the dataset VCC2018, each annotator scored an average of 226 speech. As each annotator annotates multiple speech in succession, the scores given by some annotators may be relative scores after comparison (e.g., the first or first few utterances scored by the annotator may be used as a benchmark). Moreover, compared with predicting the specific MOS score values of the speech samples, ranking the quality of speech samples has more practical application value and is often more difficult when speech samples have close MOS scores. Many previous works Cooper et al. (2022) have raised the problem of generalization, and the performance will be significantly degraded when facing the out-of-distribution (OOD) problems. Motivated by the above points, we propose a MOS prediction model based on pairwise comparison (MOSPC). Our contributions can be summarized as follows:
-
•
We propose a general framework for MOS prediction based on pair comparison, which forces the model to pay more attention to correctly rank the quality of two speech samples. To verify that MOSPC contributes to speech quality ranking, we test the ranking accuracy on the validation set on fine-grained MOS score segments. Then we utilize the C-Mixup algorithm to enhance the performance of generalization on BVCC.
-
•
Our proposed framework outperforms the baselines on BVCC and VCC2018 on most of the correlation coefficient metrics, especially on the metric KTAU related to quality ranking. And our framework surpasses the strong baseline in ranking accuracy on each fine-grained segment. These results indicate that our framework contributes to improving ranking accuracy. The model trained with VCC2018 and BVCC outperforms baselines on the OOD datasets VCC2016 and BC2019 in zero-shot experiments respectively. In addition, we analyze the performance of our model for fine-grained OOD categories, such as unseen-system, unseen-listener and unseen-speaker.
2 Related Work
A classic work in MOS prediction task is MOSNET Lo et al. (2019), which adopts the model structure of CNN-BiLSTM and proposes a loss function combining frame-level loss and utterance-level loss. Due to the need for manual annotation, few data can be used in the MOS prediction task. To reduce data waste, MBNET Leng et al. (2021) proposes a MOS predictor consisting of a meanNet and a biasNet. LDNET Huang et al. (2022) observed that MBNet removes biasNet at inference and only retains meanNet, which is inefficient. Therefore, LDNET improves MBNET by adopting an encoder-decoder structure to reduce the waste of parameters. DDOS Tseng et al. (2022) proposes to eliminate the domain mismatch between self-supervised learning (ssl) model and MOS prediction data, and adds score distribution of each utterance to model learning. UTMOS Saeki et al. (2022) is based on ensemble learning of strong and weak learners. Fusion-SSL Yang et al. (2022) uses late fusion, and fuses the results of ssl models to predict MOS value. Cooper et al. (2022) makes a analysis of the OOD problem of MOS prediction. The OOD problems in MOS prediction mainly include unseen-system, unseen-listener, unseen-speaker in the test and validation sets.
3 Method
In this section, we will introduce the overall structure of MOSPC and the implementation of pairwise comparison, as well as the C-Mixup algorithm used to enhance generalization performance.
3.1 Preliminary
Given a dataset including speech samples, denoted as , and denote the th speech sample and its ground truth. We denote the th ssl model as , , then the predicted MOS of the th ssl model for input can be represented as . represents the fusion model, which consists of ssl models and a fusion layer. denotes the predicted MOS made by the fusion model.
VCC2018 | BVCC | |||||||||||||||
utterance-level | system-level | utterance-level | system-level | |||||||||||||
MSE | LCC | SRCC | KTAU | MSE | LCC | SRCC | KTAU | MSE | LCC | SRCC | KTAU | MSE | LCC | SRCC | KTAU | |
MOSNET | 0.538 | 0.643 | 0.589 | - | 0.084 | 0.957 | 0.888 | - | 0.816 | 0.294 | 0.263 | - | 0.563 | 0.261 | 0.266 | - |
LDNET | 0.441 | 0.664 | 0.626 | 0.465 | 0.022 | 0.978 | 0.932 | 0.825 | 0.338 | 0.774 | 0.773 | 0.582 | 0.139 | 0.896 | 0.893 | 0.714 |
MBNET | 0.426 | 0.680 | 0.647 | - | 0.029 | 0.977 | 0.949 | - | 0.433 | 0.727 | 0.753 | 0.564 | 0.228 | 0.844 | 0.870 | 0.685 |
Fusion-SSL | 0.359 | 0.740 | 0.711 | 0.542 | 0.018 | 0.991 | 0.984 | 0.914 | 0.156 | 0.902 | 0.901 | 0.735 | 0.051 | 0.960 | 0.962 | 0.848 |
MOSPC | 0.352 | 0.748 | 0.721 | 0.551 | 0.020 | 0.993 | 0.988 | 0.938 | 0.148 | 0.906 | 0.906 | 0.742 | 0.054 | 0.960 | 0.962 | 0.841 |
3.2 MOSPC
3.2.1 Fusion Model
Our model is based on Fusion-SSL Yang et al. (2022). The overall model structure is shown in Figure 1. The fusion model mainly consists of 7 ssl models: wav2vec_small, wav2vec_large, hubert_base, wav2vec_large(lv60), wavlm_base, wavlm_base+, wavlm_large and a fusion layer. The fusion layer is a fully connected layer. During inference, speech is fed to ssl model separately, and the MOS values are obtained. Then the MOS values are concatenated and fed into the fusion layer to predict MOS value . During training, we leverage pairwise comparison to force the model to pay more attention to the relative quality of speech.
3.2.2 Training in Stages
Pair Comparison
Our proposed training process is shown in Figure 2. We dynamically make pairs in each batch and constrain each speech sample to form at most two pairs in order to prevent overfitting. The speech samples and are input into the ssl model respectively, then MOS scores and are predicted, and the loss function is calculated jointly. All ssl models are trained in such a pair comparison manner. Loss function consists of three parts: the relative ranking loss and the L1 loss of two speech samples denoted by and respectively:
(1) |
where is a hyperparameter, the model learns to predict MOS scores by optimizing ,, and learns to rank two audios by optimizing Burges et al. (2005). Refer to Appendix A for more details of .
C-Mixup
We observed degradation in generalization performance in the experiments on the BVCC dataset. Therefore, after experiments in Section 5.2 and 5.3, for each ssl model trained on BVCC, we adopt C-Mixup Yao et al. (2022); Cheng et al. (2023) to enhance the generalization performance. C-Mixup proportionally combines in-set samples in pairs to construct pseudo-out-of-set samples to improve the generalization performance of the model. Refer to Appendix B for details of C-Mixup algorithm. To distinguish the model from the one trained without C-Mixup, we named the model trained with C-Mixup as MOSPC-C.
4 Dataset
The datasets adopted in this paper include main track data and out-of-domain data. Main track data include VCC2018 and BVCC, and out-of-domain data include VCC2016, BC2019 and ASV2019. See Appendix C for details of datasets.
5 Experiments and Discussion
In this section, we will compare the performance of MOSPC with the baselines Lo et al. (2019); Leng et al. (2021); Huang et al. (2022) and strong baseline Fusion-SSL Yang et al. (2022) on the datasets BVCC and VCC2018. We test the generalization performance on BC2019 and VCC2016. We also list the ranking accuracy of fine-grained MOS segments.
5.1 Experiment Settings
We leverage a fusion layer and ssl models to form the overall model. Each ssl model was trained in the pair comparison manner with SGD optimizer for 1000 epochs. We applied early stopping based on the L1 loss of the validation set with 20 epochs patience, and set the learning rate to 1e-4, batch size of 8. The hyperparameter was set to be 0.6. After training on the BVCC in a pair comparison manner, we also use the C-Mixup algorithm to enhance the generalization performance of the model. When trained with the C-Mixup algorithm, We set the bandwidth to be 1.0, to be 2.0. We implemented our models in Fairseq Ott et al. (2019). All experiments were performed on 7 32GB GPUs.
5.2 MOS Prediction Results
The left side of Table 1 shows the results of MOSPC and baselines on VCC2018. MOSPC outperforms baselines in all correlation coefficient metrics of utterance-level. At the system-level, all correlation coefficient metrics outperform baselines except the MSE, which was slightly higher by 0.002. The remarkable thing is that the KTAU of system-level surpasses the baselines significantly. KTAU is a correlation coefficient metric used to indicate the ranking correlation between the predicted value and the ground truth. These results indicate that our framework contributes to the ranking correctness improvement, which is in line with our motivation.
The right side of Table 1 shows the results of our proposed MOSPC and baselines on BVCC. The results show that our model outperforms previous works on all correlation coefficient metrics at utterance-level, especially on KTAU. At the system-level our framework matches the strong baseline performance on LCC and SRCC. As there are unseen-system samples in the BVCC validation set, the performance of system-level will be affected by the samples from unseen systems. These results also imply that the pair-wise training may impair the generalization performance. To solve this problem, we adopted C-Mixup algorithm to improve the generalization performance of our model trained on BVCC.
5.3 Ranking Accuracy on Fine-Grained MOS Segments
To prove that our proposed framework contributes to the improvement of speech ranking accuracy, we analyzed the ranking accuracy of speech quality on fine-grained MOS score segments. As shown in Table 2, on BVCC and VCC2018, we divided the ground truth 1-5 into four segments with a score interval of 1. On each fine-grained MOS segments and the overall MOS segment 1-5, we calculated the ranking accuracy on each speech pair with ground truth . This is because, from our motivation, we pay more attention to whether our proposed framework can accurately rank speech with different but close score values.
BVCC | |||||
1-2 | 2-3 | 3-4 | 4-5 | 1-5 | |
Fusion-SSL | 0.728 | 0.724 | 0.739 | 0.675 | 0.778 |
MOSPC | 0.731 | 0.737 | 0.742 | 0.679 | 0.787 |
VCC2018 | |||||
Fusion-SSL | 0.482 | 0.469 | 0.515 | 0.509 | 0.493 |
MOSPC | 0.489 | 0.473 | 0.517 | 0.514 | 0.494 |
The top half of Table 2 shows the fine-grained MOS segments ranking results on the validation set of BVCC. The bottom half of Table 2 shows the fine-grained MOS segments ranking results on the validation set of VCC2018. The result shows that our proposed framework outperforms the strong baseline in ranking accuracy on each segment on both BVCC and VCC2018. These results indicate that our framework contributes to improving the ranking accuracy of speech samples with different but close MOS scores.
5.4 OOD Experiments
We first analyze the generalization performance of models trained with VCC2018 on VCC2016. As shown in Table 3, since VCC2016 only has system-level labels, we only present the system-level results. Our proposed framework outperforms previous works in all metrics, and the improvement is also significant in the KTAU metric, which again proves that our proposed framework contributes to correctly ranking the relative quality of speech.
VCC2016 | ||||
---|---|---|---|---|
system-level | ||||
MSE | LCC | SRCC | KTAU | |
MBNET | 0.207 | 0.931 | 0.906 | - |
LDNET | 0.215 | 0.939 | 0.896 | 0.768 |
Fusion-SSL | 0.209 | 0.971 | 0.889 | 0.768 |
MOSPC | 0.121 | 0.983 | 0.935 | 0.832 |
As mentioned before, from the experiments on the BVCC validation set we found that the pair-wise training method may lead to a decrease in generalization performance, so we leveraged the C-Mixup algorithm to improve the generalization performance on the BVCC after experiments in section 5.2 and 5.3. Table 4 lists the zero-shot results of BC2019. The zero-shot results indicate that after training with C-Mixup, the generalization performance improved significantly, and the robustness to the unseen-system and multi-languages OOD challenges is also improved.
BC2019 | ||||||
---|---|---|---|---|---|---|
utterance-level | system-level | |||||
LCC | SRCC | KTAU | LCC | SRCC | KTAU | |
LDNET | 0.384 | 0.365 | 0.252 | 0.500 | 0.473 | 0.354 |
DDOS | 0.678 | 0.694 | 0.502 | 0.766 | 0.797 | 0.637 |
Fusion-SSL | 0.718 | 0.642 | 0.469 | 0.803 | 0.792 | 0.601 |
MOSPC | 0.704 | 0.709 | 0.523 | 0.731 | 0.778 | 0.594 |
MOSPC-C | 0.756 | 0.711 | 0.521 | 0.816 | 0.851 | 0.667 |
unseen-speaker | unseen-system | unseen-listener | ||
---|---|---|---|---|
ASV2019 | ASV2019 | BC2019 | ASV2019 | |
Fusion-ssl | 1.104±1.641 | 1.114±1.707 | 0.191±0.225 | 1.032±1.558 |
MOSPC | 1.098±1.602 | 1.124±1.690 | 0.189±0.213 | 1.041±1.572 |
MOSPC-C | 1.089±1.587 | 1.103±1.673 | 0.179±0.217 | 1.030±1.547 |
We followed Cooper et al. (2022) analyzing the performance of our model on the fine-grained OOD categories of unseen-system, unseen-listener and unseen-speaker with ASV2019 and BC2019. We first adopted ASV2019 and BC2019 to fine-tune the model trained on BVCC respectively. As shown in table 5, we report the mean and standard deviations of squared errors for the unseen categories on utterance-level. The results indicate that our proposed method performs better on the category unseen-listener than on unseen-speaker and unseen-system.
6 Conclusion
This paper proposes a general framework for MOS prediction based on pairwise comparisons (MOSPC) to solve the problem that it is difficult for MOS prediction models to correctly rank speech quality when the MOS scores are close. The main track experiment results show that MOSPC outperforms baselines on most of the correlation coefficient metrics, especially on the metric KTAU related to speech quality ranking. Moreover, MOSPC surpasses the strong baseline in ranking accuracy on each fine-grained segment. These results indicate that training in a pair comparison manner contributes to improving ranking accuracy. We leverage C-Mixup algorithm to enhance the generalization performance. On the OOD datasets VCC2016 and BC2019, our method outperforms baselines on all metrics. We also analyze the performance on fine-grained OOD categories. Our method performs better for the unseen-listener OOD category than for the unseen-speaker and unseen-system OOD categories.
7 Limitation
MOSPC can improve ranking accuracy on each fine-grained MOS score segment, but at the same time, the training method based on pair comparison may impair the generalization performance. As there are unseen-systems in the BVCC validation set, the system-level results of BVCC are affected by the generalization performance degradation. We introduced the C-Mixup algorithm to enhance the generalization performance, which increased the complexity of the experiment to some extent.
References
- Burges et al. (2005) Chris Burges, Tal Shaked, Erin Renshaw, Ari Lazier, Matt Deeds, Nicole Hamilton, and Greg Hullender. 2005. Learning to rank using gradient descent. In Proceedings of the 22nd international conference on Machine learning, pages 89–96.
- Cheng et al. (2023) Xuxin Cheng, Qianqian Dong, Fengpeng Yue, Tom Ko, Mingxuan Wang, and Yuexian Zou. 2023. M 3 st: Mix at three levels for speech translation. In ICASSP 2023-2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pages 1–5. IEEE.
- Cooper et al. (2022) Erica Cooper, Wen-Chin Huang, Tomoki Toda, and Junichi Yamagishi. 2022. Generalization ability of mos prediction networks. In ICASSP 2022-2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pages 8442–8446. IEEE.
- Das et al. (2020) Rohan Kumar Das, Tomi Kinnunen, Wen-Chin Huang, Zhenhua Ling, Junichi Yamagishi, Yi Zhao, Xiaohai Tian, and Tomoki Toda. 2020. Predictions of subjective ratings and spoofing assessments of voice conversion challenge 2020 submissions. arXiv preprint arXiv:2009.03554.
- Huang et al. (2022) Wen-Chin Huang, Erica Cooper, Junichi Yamagishi, and Tomoki Toda. 2022. Ldnet: Unified listener dependent modeling in mos prediction for synthetic speech. In ICASSP 2022-2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pages 896–900. IEEE.
- Kim (2005) Doh-Suk Kim. 2005. Anique: An auditory model for single-ended speech quality estimation. IEEE Transactions on Speech and Audio Processing, 13(5):821–831.
- King et al. (2008) Simon King, Robert AJ Clark, Catherine Mayo, and Vasilis Karaiskos. 2008. The blizzard challenge 2008.
- King and Karaiskos (2014) Simon King and Vasilis Karaiskos. 2014. The blizzard challenge 2013.
- Kinga and Karaiskosb (2009) Simon Kinga and Vasilis Karaiskosb. 2009. The blizzard challenge 2009. In The Blizzard Challenge 2009 Workshop.
- Kinga and Karaiskosb (2010) Simon Kinga and Vasilis Karaiskosb. 2010. The blizzard challenge 2010. In The Blizzard Challenge 2010 Workshop.
- Kinga and Karaiskosb (2011) Simon Kinga and Vasilis Karaiskosb. 2011. The blizzard challenge 2011. In The Blizzard Challenge 2011 Workshop.
- Kinga and Karaiskosb (2016) Simon Kinga and Vasilis Karaiskosb. 2016. The blizzard challenge 2016. In The Blizzard Challenge 2011 Workshop.
- Kubichek (1993) Robert Kubichek. 1993. Mel-cepstral distance measure for objective speech quality assessment. In Proceedings of IEEE pacific rim conference on communications computers and signal processing, volume 1, pages 125–128. IEEE.
- Leng et al. (2021) Yichong Leng, Xu Tan, Sheng Zhao, Frank Soong, Xiang-Yang Li, and Tao Qin. 2021. Mbnet: Mos prediction for synthesized speech with mean-bias network. In ICASSP 2021-2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pages 391–395. IEEE.
- Lo et al. (2019) Chen-Chou Lo, Szu-Wei Fu, Wen-Chin Huang, Xin Wang, Junichi Yamagishi, Yu Tsao, and Hsin-Min Wang. 2019. Mosnet: Deep learning based objective assessment for voice conversion. arXiv preprint arXiv:1904.08352.
- Lorenzo-Trueba et al. (2018a) Jaime Lorenzo-Trueba, Junichi Yamagishi, Tomoki Toda, Daisuke Saito, Fernando Villavicencio, Tomi Kinnunen, and Zhenhua Ling. 2018a. The voice conversion challenge 2018: Promoting development of parallel and nonparallel methods. arXiv preprint arXiv:1804.04262.
- Lorenzo-Trueba et al. (2018b) Jaime Lorenzo-Trueba, Junichi Yamagishi, Tomoki Toda, Daisuke Saito, Fernando Villavicencio, Tomi Kinnunen, and Zhenhua Ling. 2018b. The voice conversion challenge 2018: Promoting development of parallel and nonparallel methods. arXiv preprint arXiv:1804.04262.
- Malfait et al. (2006) Ludovic Malfait, Jens Berger, and Martin Kastner. 2006. P. 563—the itu-t standard for single-ended speech quality assessment. IEEE Transactions on Audio, Speech, and Language Processing, 14(6):1924–1934.
- Ott et al. (2019) Myle Ott, Sergey Edunov, Alexei Baevski, Angela Fan, Sam Gross, Nathan Ng, David Grangier, and Michael Auli. 2019. fairseq: A fast, extensible toolkit for sequence modeling. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics (Demonstrations), pages 48–53.
- Saeki et al. (2022) Takaaki Saeki, Detai Xin, Wataru Nakata, Tomoki Koriyama, Shinnosuke Takamichi, and Hiroshi Saruwatari. 2022. Utmos: Utokyo-sarulab system for voicemos challenge 2022. arXiv preprint arXiv:2204.02152.
- Toda et al. (2016) Tomoki Toda, Ling-Hui Chen, Daisuke Saito, Fernando Villavicencio, Mirjam Wester, Zhizheng Wu, and Junichi Yamagishi. 2016. The voice conversion challenge 2016. In Interspeech, pages 1632–1636.
- Todisco et al. (2019) Massimiliano Todisco, Xin Wang, Ville Vestman, Md. Sahidullah, Héctor Delgado, Andreas Nautsch, Junichi Yamagishi, Nicholas Evans, Tomi H Kinnunen, and Kong Aik Lee. 2019. ASVspoof 2019: future horizons in spoofed and fake audio detection. In Proc. Interspeech, pages 1008–1012.
- Tseng et al. (2022) Wei-Cheng Tseng, Wei-Tsung Kao, and Hung-yi Lee. 2022. Ddos: A mos prediction framework utilizing domain adaptive pre-training and distribution of opinion scores. arXiv preprint arXiv:2204.03219.
- Wang et al. (2020) Xin Wang, Junichi Yamagishi, Massimiliano Todisco, Héctor Delgado, Andreas Nautsch, Nicholas Evans, Md Sahidullah, Ville Vestman, Tomi Kinnunen, Kong Aik Lee, Lauri Juvela, Paavo Alku, Yu-Huai Peng, Hsin-Te Hwang, Yu Tsao, Hsin-Min Wang, Sébastien Le Maguer, Markus Becker, Fergus Henderson, Rob Clark, Yu Zhang, Quan Wang, Ye Jia, Kai Onuma, Koji Mushika, Takashi Kaneda, Yuan Jiang, Li-Juan Liu, Yi-Chiao Wu, Wen-Chin Huang, Tomoki Toda, Kou Tanaka, Hirokazu Kameoka, Ingmar Steiner, Driss Matrouf, Jean-François Bonastre, Avashna Govender, Srikanth Ronanki, Jing-Xuan Zhang, and Zhen-Hua Ling. 2020. ASVspoof 2019: a large-scale public database of synthesized, converted and replayed speech. Computer Speech & Language, page 101114.
- Watanabe et al. (2018) Shinji Watanabe, Takaaki Hori, Shigeki Karita, Tomoki Hayashi, Jiro Nishitoba, Yuya Unno, Nelson Enrique Yalta Soplin, Jahn Heymann, Matthew Wiesner, Nanxin Chen, et al. 2018. Espnet: End-to-end speech processing toolkit. arXiv preprint arXiv:1804.00015.
- Wester et al. (2016a) Mirjam Wester, Zhizheng Wu, and Junichi Yamagishi. 2016a. Analysis of the voice conversion challenge 2016 evaluation results. In Interspeech, pages 1637–1641.
- Wester et al. (2016b) Mirjam Wester, Zhizheng Wu, and Junichi Yamagishi. 2016b. Analysis of the voice conversion challenge 2016 evaluation results. In Interspeech, pages 1637–1641.
- Yang et al. (2022) Zhengdong Yang, Wangjin Zhou, Chenhui Chu, Sheng Li, Raj Dabre, Raphael Rubino, and Yi Zhao. 2022. Fusion of self-supervised learned models for mos prediction. arXiv preprint arXiv:2204.04855.
- Yao et al. (2022) Huaxiu Yao, Yiping Wang, Linjun Zhang, James Zou, and Chelsea Finn. 2022. C-mixup: Improving generalization in regression. arXiv preprint arXiv:2210.05775.
- Zhao et al. (2020) Yi Zhao, Wen-Chin Huang, Xiaohai Tian, Junichi Yamagishi, Rohan Kumar Das, Tomi Kinnunen, Zhenhua Ling, and Tomoki Toda. 2020. Voice conversion challenge 2020: Intra-lingual semi-parallel and cross-lingual voice conversion. arXiv preprint arXiv:2008.12527.
Appendix A Details of the Relative Ranking Loss
was introduced by rankNetBurges et al. (2005). is similar in form to cross entropy loss:
(2) |
where maps the outputs , into probability via a logistic function:
(3) |
The value of depends on the ground truths of two speech samples.
(4) |
Appendix B C-Mixup
For ssl model and input speech sample , we need to sample another instance from the training set. C-Mixup first constructs a sampling probability distribution based on a symmetric Gaussian kernel for each audio sample :
(5) |
where represents the distance between and , and represents the bandwidth which is a hyperparameter. Subsequently, these conditional probabilities are normalized into a probability mass function that sums to one, and another sample is selected by sampling through the probability mass function. Figure 3 illustrates the training process of C-Mixup. Each ssl model in this work contains two parts: feature extractor and encoder. and are fed into the feature extractor respectively to obtain embedding and . Then and are proportionally combined to construct the embedding of pseudo-out-of-set sample:
(6) |
where , and is the parameter of the Beta distribution. is a hyperparameter. The remaining models take the pseudo-out-of-set embedding as input to predict MOS score , and compute the L1 loss with . is constructed in the same way as :
(7) |
consistent with the main track, each ssl model is trained with C-Mixup algorithm separately.
Appendix C Details of Dataset
C.1 Main Track Data
C.1.1 VCC2018
Samples in VCC2018 were all sampled from Voice Conversion Challenge 2018Lorenzo-Trueba et al. (2018a), including English speech samples synthesized by systems. A total of professional annotators participated in the speech labeling. Each speech was scored by four annotators, and the four integer scores were averaged as the label. For the sake of comparison, We split VCC2018 into training sets with a size of , validation sets with a size of and test sets with a size of .
C.1.2 BVCC
BVCC integrates data from multiple synthetic speech competitions, including Blizzard ChallengeKing et al. (2008); Kinga and Karaiskosb (2009, 2010, 2011); King and Karaiskos (2014); Kinga and Karaiskosb (2016), the Voice Conversion ChallengeToda et al. (2016); Wester et al. (2016b); Lorenzo-Trueba et al. (2018b); Zhao et al. (2020); Das et al. (2020) and publicly-available samples from systems implemented in ESPnetWatanabe et al. (2018). BVCC includes a total of 7106 English speech samples submitted by 187 systems. We split BVCC into training, validation and test sets with a rate of , and . Each speech was scored by eight annotators, and the eight integer scores were averaged as the label. Unlike VCC2018, BVCC has samples from unseen systems in its validation set.
C.2 Out-of-domain Data
C.2.1 VCC2016
In order to compare with previous works, we adopt VCC2016 to test the OOD performance of models trained with VCC2018. VCC2016 includes speech samples synthesized by systems. VCC2016 has only system-level labels and without utterance-level labels.
C.2.2 BC2019
We adopt BC2019 to test the OOD performance of models trained with BVCC. Samples in BC2019 are all sampled from Blizzard Challenge 2019, and are Chinese TTS synthesized speech rated by Chinese native speakers. Since all samples of BVCC are in English, BC2019 can be used as a cross-language OOD case to test the generalization performance of models trained with BVCC. BC2019 has provided 136 labeled samples for training, 136 samples for validation, and additional 540 unlabeled data.
C.3 ASV2019
We follow Cooper et al. (2022) utilizing ASV2019Wang et al. (2020); Todisco et al. (2019) to analyze the performance of our model on fine-grained OOD experiments. Samples in ASV2019 are all in English and sampled from the human assessment results data on the ASVspoof2019 database LA scenario. As scores in human assessment results data are distributed from 0 to 9, We linearly project the scores to 1-5.