This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

PoLyScriber: Integrated Fine-tuning of Extractor and Lyrics Transcriber for Polyphonic Music

Xiaoxue Gao,  Chitralekha Gupta,  Haizhou Li Manuscript received October 1, 2022; revised March 1, 2023; accepted May 2, 2023. This work is supported by the Agency for Science, Technology and Research (A*STAR) under its AME Programmatic Funding Scheme (Project No. A18A2b0046), and IAF, A*STAR, SOITEC, NXP and National University of Singapore under FD-fAbrICS: Joint Lab for FD-SOI Always-on Intelligent & Connected Systems (Award I2001E0053).Xiaoxue Gao is with the Department of Electrical and Computer Engineering, National University of Singapore, Singapore 117583 (e-mail: [email protected]).Chitralekha Gupta is with the School of Computing, National University of Singapore, Singapore 117417 (e-mail: [email protected]).Haizhou Li is with Shenzhen Research Institute of Big Data, School of Data Science, The Chinese University of Hong Kong, Shenzhen, China 518172, the Department of Electrical and Computer Engineering, National University of Singapore, Singapore 117583 (e-mail: [email protected]).
Abstract

Lyrics transcription of polyphonic music is challenging as the background music affects lyrics intelligibility. Typically, lyrics transcription can be performed by a two-step pipeline, i.e. a singing vocal extraction front end, followed by a lyrics transcriber back end, where the front end and back end are trained separately. Such a two-step pipeline suffers from both imperfect vocal extraction and mismatch between front end and back end. In this work, we propose a novel end-to-end integrated fine-tuning framework, that we call PoLyScriber, to globally optimize the vocal extractor front end and lyrics transcriber back end for lyrics transcription in polyphonic music. The experimental results show that our proposed PoLyScriber achieves substantial improvements over the existing approaches on publicly available test datasets.

Index Terms:
Lyrics transcription in polyphonic music, music information retrieval, integrated fine-tuning, vocal extraction

I Introduction

Lyrics constitute the fundamental textual component of singing voice in music for the emotional perception of the song. Lyrics also helps in foreign language learning [1]. Music analysis has attracted great attention recently including audio-visual music source separation, localization [2, 3, 4] and music generation [5, 6]. Being able to understand the sung text may well contribute to listeners’ enjoyment of music [7]. Lyrics transcription seeks to recognize the sung lyrics from singing vocals in the presence of instrumental music accompaniment. Various applications, such as automatic generation of karaoke lyrical content, music video subtitling [8] and query-by-singing [9] can benefit from automatic lyrics transcription.

In lyrics transcription, we encounter many technical challenges. A prior study [10] shows that lyrics transcription in polyphonic music is challenging even for professional musicians, as the intelligibility is affected by many factors including complex structure of singing, diverse polyphonic music and different environmental conditions. Specifically, singing shows a higher degree of pronunciation and prosody variation than speech [11, 12, 13, 14], therefore lyrics transcription is more challenging than automatic speech recognition (ASR). Moreover, background music accompaniment interferes with sung voice. The accompaniment typically adds an intricate and structured source of musical information to the singing vocals, and often influences the lyrics intelligibility of the singing vocal [15, 16]. Thus, any detrimental effects of singing to lyrics intelligibility are likely to be exacerbated by background music distraction.

To address the problem of automatic lyrics transcription in polyphonic music, there are mainly two broad approaches, 1) direct modeling (DM) approach which takes the polyphonic audio, i.e. singing vocals + background music as model input directly to transcribe lyrics [17, 18], and 2) extraction-transcription approach which extracts singing vocals first through a trained source separation model, and then transcribes the lyrics of the extracted vocals [19, 20, 21, 22].

In direct modeling approaches, acoustic models are trained directly on the polyphonic music, a mixture of vocals and music accompaniments. For example, Stoller et al. [17] developed a system based on a Wave-U-Net architecture to predict character probabilities directly from raw audio. This system performed well for the task of lyrics-to-audio alignment, however, showed a high word error rate (WER) in lyrics transcription. Music-informed acoustic modelling that incorporated music genre-specific information are proposed [18, 23]. The study in [18] suggested that lyrics acoustic models can benefit from music genre knowledge of the background music but it requires an additional genre extraction step, separately. Moreover, background music is difficult to model explicitly due to complicated rhythmic and harmonic structure, diverse music genres and compositional styles [10]. The effect of interference caused by the background music has not been well-studied under direct modeling approaches.

Humans are able to recognize lyrical words of a song by paying more attention to the singing vocal than the background music. The extraction-transcription approach [20, 24, 25] attempts to emulate such as human listening process through a two-step pipeline in which, first, the singing vocal is extracted from the polyphonic input signal, then the lyrics are transcribed from the extracted singing vocal. The two-step pipeline suffers from both imperfect vocal extraction and mismatch between the vocal extraction front end and the lyrics decoding back end [19, 18]. In the two-step pipeline, the front end is not optimized for lyrics transcription, and the mismatch is partly because front end and back end are trained separately with different objective functions.

In this paper, we propose an end-to-end (E2E) integrated fine-tuning solution, PoLyScriber, that integrate vocal extraction with lyrics transcription in a single network. This is a departure from the previous studies [19, 18, 20, 24, 25], where extraction and transcription are trained separately with mismatch problem. To the best of our knowledge, this work presents the first study on an integrated extraction-transcription solution to automatic lyric transcription, where the integrated fine-tuning is designed to fine-tune an integrated framework that consists of a pre-trained extractor and a pre-trained lyrics transcriber. The contributions of this paper include: 1) novel E2E neural solutions for automatic lyrics transcription that combines the two-step pipeline into one through an integrated fine-tuning framework; 2) a systematic comparative study over several solutions and provide our insightful findings that the appropriate amount of removal of background accompaniments can be achieved by the integrated fine-tuning, where its intermediate vocal outputs lie in the middle of complete music removal and complete music presence.

The rest of this paper is organized as follows. Section II presents the related work to set the stage for this study. In Section III, we present the overview of our proposed PoLyScriber. Section IV introduces the components of the proposed PoLyScriber in detail. Section V consists of the database and experimental setup. Section VI discusses the experiment results. Finally, Section VII concludes the study.

II Related Work

We discuss the related work in singing voice separation as well as lyrics transcription of solo-singing and polyphonic music to motivate this study.

II-A Singing Voice Separation

The extraction-transcription approach makes use of a singing voice separation (SVS) front end to separate singing vocals from polyphonic music. There are generally two approaches for SVS, namely waveform-based and spectrogram-based approaches [26]. The waveform-based approach directly takes time-domain polyphonic music signal as input and separately output singing voice and music signals [27, 28, 29, 30]. As the waveform-based encoder learns directly from input data, it is sensitive to the change of music content. The spectrogram-based approach applies frequency analysis in the front end. It predicts a power spectrum for each source and re-use the phase from the input mixture to synthesize two separated outputs. Open Unmix [31] and MMDenseLSTM [32] are examples of spectrogram-based approaches, that show good results in the SiSEC 2018 evaluation [33] on the MusDB dataset [34]. Spleeter [35] is another spectrogram-based system that has shown a strong performance and has been widely adopted by the digital music industry, but it is now outperformed by more advanced spectrogram domain architectures such as D3Net [36] and ByteMSS [37], even though Spleeter is trained on much more data from Bean dataset [38] with nearly 25,000 songs.

The spectrogram-based separation methods often suffer from incorrect phase reconstruction. Kong et al. [37] proposed ByteMSS to estimate the phase by estimating complex ideal ratio masks where they decouple the estimation of these masks into magnitude and phase estimations based on a deep residual U-Net model. This system achieves the state-of-the-art (SOTA) singing voice separation result of SDR with 8.98 dB on the MUSDB18 test dataset. This opens doors to many other closely related MIR tasks, such as music transcription, singing melody transcription and lyrics transcription.

In this paper, we explore end-to-end neural solutions to lyrics transcription that benefits from the best of both direct modeling and extraction-transcription studies. We explore the use of ByteMSS network for singing voice extraction.

II-B Lyrics Transcription of Vocal and Polyphonic Music

The singing vocal carries the lyrics of a song, therefore lyrics transcription of polyphonic music can benefit a lot from the recent advances in the field of lyrics transcription of solo-singing [26, 39].

II-B1 Lyrics Transcription of Solo-singing

Lyrics transcription of solo-singing is typically performed by a speech recognition engine, such as Kaldi [40]. Gupta et al. developed a DNN-HMM using an early version of DAMP dataset and reported 29.65% word error rate (WER) with duration-adjusted phonetic lexicon. More recently, Dabike et al. [41] created a cleaner version of DAMP dataset and constructed a factorized Time-Delay Neural Network (TDNN-F) using Kaldi [40] with WER of 19.60%. Demirel et al. [25] further incorporated convolutional and self-attention layers to TDNN-F and achieved WER 14.96%. Our recent work [42] introduced an end-to-end Transformer-based framework along with RNN-based language model and is shown to be competitive with current approaches. This provides possibility and flexibility for lyrics transcription in polyphonic music where we would need a pre-trained solo-singing model to initialize the acoustic model for the integrated fine-tuning process.

Refer to caption
Figure 1: (a) The overview network architecture of the proposed integrated fine-tuning approach (2) PoLyScriber with its two variants (1)(3). The detailed network architecture of (2) PoLyScriber includes (b) singing vocal extractor and (c) singing lyrics transcriber. PoLyScriber-NoAug (1) and PoLyScriber-L (3) have the same network architecture as PoLyScriber, while PoLyScriber-NoAug (1) does not employ data augmentation and PoLyScriber-L (3) has the additional extraction loss.

II-B2 From Solo-Singing to Polyphonic Music

Studies show that acoustic models trained on solo-singing data [43, 44, 45, 46, 47, 48, 49, 41] performs poorly on polyphonic music test data [50, 51, 24]. One way to adapt a solo-singing acoustic model towards polyphonic music is through domain adaptation [52]. It is found that an acoustic model adapted on polyphonic music data outperforms that adapted on extracted vocals. This suggests that singing voice extraction or separation introduces unwanted artifacts, that adversely affect the lyrics transcription performance. Despite much progress, the lyrics transcription of polyphonic music remains to be improved. As singing vocals and background music are highly correlated in polyphonic music, resulting in overlapping frequency components, lyrics transcription of polyphonic music is more challenging than that of solo-singing.

II-B3 Towards End-to-End Singing Decoder

Many lyrics transcription systems [20, 24, 25, 17, 18, 53, 54] are based on hybrid modular architecture, that consists of acoustic, lexical and language models [40], each with a different objective, thereby suffering from disjoint training issue. Studies show that end-to-end (E2E) systems bring us many advantages [55, 56, 57, 58, 59, 60] including its simpliciy without the need of elaborate controlling of GMM, HMM and neural network models, its independent capability of lexicon and its flexibility to incorporate other models. E2E model only needs one single neural network with one objective function to optimize for the lyrics transcription task. We reported an E2E transformer-based system for lyrics transcription of polyphonic music [42] that outperforms other approaches in the literature. It consists of a transformer based encoder-decoder framework with a multi-head attention that implicitly learns the global time contextual dependency of the singing vocal utterance, beyond just the current frame. We adopt this singing decoder [42] as our network backbone.

III Overview of PoLyScriber

III-A Motivation

TABLE I: Comparison of lyrics recognition (WER%) performances under different conditions of polyphonic music testsets on the Solo-singing transcriber (SoloM) model as detailed in Section VI-B. The extracted vocal and bgm are obtained by sRes Extractor as described in Section V-B1 from polyphonic music testsets in Section V-A1. The lower the SDR is, the higher-level energy of the bgm are added.
Decoding input audio types to SoloM Hansen Jamendo Mauch
extracted vocal 66.76 62.48 58.28
extracted vocal + bgm (SDR = 20dB) 59.33 59.82 55.46
extracted vocal + bgm (SDR = 15 dB) 58.67 58.72 54.67
extracted vocal + bgm (SDR = 5 dB) 56.09 59.24 55.38
extracted vocal + bgm (SDR = 10 dB) 57.24 58.49 55.50
extracted vocal + bgm (SDR = 0 dB) 59.96 61.92 61.62
extracted vocal + bgm (SDR = -5 dB) 67.77 69.16 74.20
extracted vocal + bgm (SDR = -10 dB) 84.03 81.94 89.02
extracted vocal + bgm (SDR = -15 dB) 93.37 90.70 95.66
extracted vocal + bgm (SDR = -20 dB) 97.10 94.80 96.70
Original polyphonic tracks 65.54 65.65 69.43

Since singing vocals are highly correlated and overlapped with the background music in polyphonic music, it is difficult to achieve a perfect singing voice extraction. In many studies, singing extractors and singing acoustic modeling are two steps in a pipeline, where each of these modules is independently trained [20, 24, 25] and extracted vocals often suffer from distortions in the extractor. On the other hand, [18, 23] have shown promising results where the vocal extraction step is avoided and the vocals along with the accompanying music are modeled together for lyrics recognition (i.e. direct modeling). However, the music accompaniments can be loud which can affect lyrics intelligibility and hence the performance of the lyrics recognition system [18, 23, 17].

We hypothesize that for lyrics recognition, the front end should be able to suppress loud background accompaniments while not causing vocal distortion or vocal removal. One way to test this theory is to add back the extracted background accompaniments (bgm) to the extracted vocal at different signal-to-distortion ratios (SDR) and pass this combined signal through a transcriber trained on singing vocals to observe its effect on the lyrics recognition performance. Indeed, we observe (Table I) that in the presence of low volume background accompaniments, lyrics recognition performance is better than the two extreme ends - extracted vocals input with completely suppressed background music, and the original vocals+music input. The reason is that extracted vocals with complete music suppression results in partial removal and distortions in the vocals while the original polyphonic music tracks have loud background music that affects lyrics recognition performance. Adding in some of the removed music also adds in the vocal parts, thus improving recognition performance.

This raises the question - How to automatically cause the ideal amount of music suppression required for lyrics recognition? To answer this question, in this work, we propose an integrated optimization strategy for the extractor and the acoustic model for the task of lyrics transcription.

III-B Integrated Fine-tuning Approach

In this work, we propose an end-to-end lyrics transcription framework, PoLyScriber, in which we holistically train the singing voice extraction module and the singing acoustic model in a single integrated network utilizing polyphonic data augmentation for polyphonic music, as shown in Fig.1. To comprehensively explore the integrated fine-tuning approach, we also investigate two different variants of the PoLyScriber. These includes transcription-oriented training (PoLyScriber-NoAug) and the inclusion of the extraction loss (PoLyScriber-L), as illustrated in Fig.1.

III-B1 PoLyScriber with Polyphonic Data Augmentation

PoLyScriber is an E2E approach utilizing polyphonic data augmentation where vocal extraction model is integrated with the lyrics transcription model and they are trained together solely towards the objective of lyrics transcription. This model uses both realistic and simulated polyphonic music data for E2E training, as shown in Fig. 1 (a) with dotted lines and the symbol (2). Having access to solo-singing data, simulated polyphonic music is created by data augmentation. We conduct polyphonic data augmentation by randomly selecting music and mixing it with solo-singing data with random signal-to-noise ratios (SNR) while training the PoLyScriber. By doing so, we expect to improve the diversity of polyphonic music data and further to be beneficial for model generalization.

III-B2 PoLyScriber-NoAug Variant with Transcription-oriented Training

If solo-singing or background music data is not available, we explore a transcription-oriented training strategy, PoLyScriber-NoAug, that does not adopt polyphonic data augmentation. PoLyScriber-NoAug has the same architecture as PoLyScriber trained solely towards the objective of lyrics transcription, while the training data used for PoLyScriber-NoAug is only realistic polyphonic music data, as explained in Fig.1 (a) with symbol (1).

III-B3 PoLyScriber-L Variant with Inclusion of Extraction Loss

On top of PoLyScriber, we also introduce a variant by additionally incorporating a vocal extraction loss along with the transcription loss to supervise the extractor front-end training on the simulated as well as real data as illustrated in Fig. 1(a) with dotted lines and the symbol (3).

III-C Network Architecture

We adopt the same network architecture for PoLyScriber and its variants. We simplify a residual-UNet vocal extraction front end [37] as a simplified Residual-UNet (sRes) and concatenate it with a transformer-based singing acoustic model together to build PoLyScriber, and globally adjust the weights in each of these modules via back-propagation.

To construct base models for integrated fine-tuning initialization, we first obtain the pre-trained singing vocal extractor and the singing acoustic model (transcriber). Specifically, the extractor front end is trained to reconstruct singing vocal using parallel polyphonic music and corresponding clean singing vocals data, and the transcriber is constructed with a joint encoder-decoder and connectionist temporal classification (CTC) architecture [60] to transcribe lyrics from clean singing vocals. Second, we use the pre-trained extractor and transcriber as the base models and further globally train the extractor and transcriber on a training dataset. The integrated fine-tuning allows the linguistic information captured by the acoustic model to flow back and inform the extraction front end. By optimizing PoLyScriber network in this way, the extractor would be able to extract features in a way that is suitable for the task of lyrics transcription.

IV Components of PoLyScriber

In this section, we describe our design of each of the components in the PoLyScriber in detail.

IV-A Singing Vocal Extractor Front End

The goal of the singing vocal extractor in PoLyScriber is to remove, or at least suppress the background music to a certain level while sustaining lyrics transcription related vocal parts. The extractor is based on the Residual-UNet [37] model to estimate complex ideal ratio masks (cIRMs) and spectrogram, modified for the purpose of integrated fine-tuning. Compared with conventional approaches that do not estimate phases of the extracted signals [61, 62, 63, 32], and often suffers from poor audio quality possibly with incorrect phase reconstruction, cIRMs based system contains phase prediction and therefore, may alleviate incorrect phase reconstruction problem. However, the direct predicting of real and imaginary parts of cIRMs suffers from being sensitive to time-domain signal shifts [64, 65, 66, 37]. Considering the superiority of cIRM for estimating phase [37], we use the extractor that is designed to decouple the estimation of cIRMs into magnitude and phase estimations. The extractor also combines the predictions of the cIRM mask and spectrogram where the spectrogram prediction is the direct magnitude prediction serving as a residual component to complement the bounded mask prediction [37]. This was one of the top performing systems in the recent Music Demixing Challenge 2021 [67] with a signal-to-distortion ratio (SDR) of 8.9 dB on a standard test dataset.

We employ a simplified version of the Residual-UNet network as our front-end sRes vocal extractor that has an encoder, intermediate layers and a decoder as shown in Fig. 1 (b). We denote the cIRM as M^\hat{M}. The input short-time Fourier transforms (STFTs) XX from the realistic polyphonic music or the simulated polyphonic music is compressed by the extractor encoder and intermediate layers into a lower dimensional descriptor and then the descriptor is re-expanded to the size of the target solo-singing STFTs SS by the extractor decoder. In the following subsections, we present the details of the vocal extractor.

IV-A1 Extractor Architecture

The extractor encoder consists of four residual encoder blocks (REBs) where each REB contains two residual convolutional blocks (RCBs) followed by an 2 ×\times 2 average pooling layer to reduce the feature map size. The real/simulated polyphonic music STFT input XX is first encoded as polyphonic representations via the extractor encoder. Each RCB consists of two convolutional layers with kernel sizes 3 ×\times 3. A residual connection is added between the input and the output of a RCB [37]. A batch normalization and a leaky ReLU non-linearity with a negative slope of 0.01 is applied before each convolutional layer inside each RCB following the pre-act residual network configuration [37].

The two intermediate convolutional blocks (ICBs) then transform the polyphonic representations into a hidden descriptor, where each ICB has the same architecture as the REB except the pooling layer. The hidden descriptor is further fed into the extractor decoder, which is symmetric to those in the extractor encoder and it contains four residual decoder blocks (RDBs). Each RDB consists of a transposed convolutional layer with a kernel size 3 ×\times 3 and stride 2 ×\times 2 to upsample feature maps, followed by two RCBs. After the extractor decoder, an additional ICB and a final convolutional layer are applied to generate four outputs (the magnitude of the cIRM |M^||\hat{M}|, direct magnitude prediction |Q^||\hat{Q}|, real part of the cIRM P^r\hat{P}_{r} and the imaginary part of the cIRM P^i\hat{P}_{i}). Therefore, the extracted singing vocal STFTs S^\hat{S} is predicted following [37] as below:

S^=|S^|cosS^+j|S^|sinS^,|S^|=relu(|M^||X|+|Q^|),S^=M^+X,cosS^=cosM^cosXsinM^sinX,sinS^=sinXcosM^+cosXsinM^,sinM^=P^r/P^r2+P^i2,cosM^=P^i/P^r2+P^i2\begin{split}\hat{S}=|\hat{S}|cos\angle{\hat{S}}+j|\hat{S}|sin\angle\hat{S},\\ |\hat{S}|=relu(|\hat{M}||X|+|\hat{Q}|),\\ \angle{\hat{S}}=\angle{\hat{M}}+\angle{X},\\ cos\angle{\hat{S}}=cos\angle{\hat{M}}cos\angle{X}-sin\angle\hat{M}sin\angle X,\\ sin\angle\hat{S}=sin\angle Xcos\angle\hat{M}+cos\angle Xsin\angle\hat{M},\\ sin\angle\hat{M}=\hat{P}_{r}/\sqrt{\hat{P}_{r}^{2}+\hat{P}_{i}^{2}},\\ cos\angle\hat{M}=\hat{P}_{i}/\sqrt{\hat{P}_{r}^{2}+\hat{P}_{i}^{2}}\end{split} (1)

As illustrated in Fig. 1 (b) and Eq. 1, the angle of cIRM M^\angle{\hat{M}} and the angle of extracted vocal STFTs S^\angle{\hat{S}} can be obtained from P^r\hat{P}_{r} and P^i\hat{P}_{i}, while |Q^||\hat{Q}| is residual component to |M^||\hat{M}| for getting extracted vocal magnitude |S^||\hat{S}|. After obtaining the STFTs of the extracted vocal, an inverse STFT is applied to obtain the extracted vocal waveform.

IV-A2 Extractor Training Objective

The sRes extractor is pre-trained on parallel polyphonic music and clean solo-singing data with a L1-loss that is computed on the waveform domain as shown below:

Ext=1T(t=1T|𝐬^(t)𝐬(t)|)\begin{split}\mathcal{L}_{\text{Ext}}=\frac{1}{T}(\sum_{t=1}^{T}|\mathbf{\hat{s}}(t)-\mathbf{s}(t)|)\end{split} (2)

where the 𝐬^(t)\mathbf{\hat{s}}(t) and 𝐬(t)\mathbf{s}(t) are the extracted singing vocal waveform and the corresponding target solo-singing waveform with tt as the discrete time index, respectively.

IV-B Singing Lyrics Transcriber

The singing lyrics transcriber uses a transformer-based lyrics transcription framework which is trained to decode input feature sequence of extracted singing vocal to output lyrical sequence, as shown in Fig. 1 (c). The transcriber encoder converts the input acoustic features to intermediate representations, and the transcriber decoder predicts lyrical tokens (we use sub-words where 5,000 sub-words are generated using byte-pair encoding (BPE) in this work) one at a time given the intermediate representations and previously predicted lyrical tokens in an auto-regressive manner.

IV-B1 Transcriber Encoder

The transcriber encoder consists of an acoustic embedding module and twelve identical sub-encoders where each sub-encoder contains a multi-head attention (MHA) [68] and a position-wise feed-forward network (FFN). The extracted vocal acoustic features FF for the transcriber are first obtained from the extracted singing vocal audio via the feature extraction block, which first performs downsampling on the audio to a sampling rate of 16KHz and then extracts 80-dim filterbank feature with a window of 25 ms, shifting 10 ms. FF is then encoded into the acoustic embedding by the acoustic embedding module using subsampling and positional encoding (PE) [68]. The acoustic embedding block contains two CNN blocks with a kernel size of 3 and a stride size of 2. The sub-encoders then transform the acoustic embedding into a hidden representation HH. Residual connection [69] and layer normalization [70] are employed inside each of the sub-encoder.

H=TranscriberEncoder(F)\centering\begin{split}H=\text{TranscriberEncoder}(F)\end{split}\@add@centering (3)

IV-B2 Transcriber Decoder

The transcriber decoder consists of a lyrical embedding module and six identical sub-decoders, where each sub-decoder has a masked MHA, a MHA and a FFN. During training, YY represents the lyrical token history that is offset right by one position, however, during run-time inference, it represents the previous predicted token history. YY is first converted to lyrics token embedding via lyrical embedding module, that consists of an embedding layer, and a positional encoding (PE) operation.

O=TranscriberDecoder(H,Y)\centering\begin{split}O=\text{TranscriberDecoder}(H,Y)\end{split}\@add@centering (4)
TABLE II: A description of polyphonic music dataset that consists of DALI and NUS collections.
Name # songs Lyrical lines Duration
Poly-train DALI-train 3,913 180,034 208.6 hours
NUS 517 264,62 27.0 hours
Poly-dev DALI-dev 100 5,356 3.9 hours
NUS 70 2,220 3.5 hours
Poly-test Hansen 10 212 0.5 hour
Jamendo 20 374 0.9 hour
Mauch 20 442 1.0 hour

The lyrics embedding is fed into the masked MHA that ensures causality, i.e. the predictions for current position only depends on the past positions. The output of the masked MHA and the acoustic encoding HH are then fed to the next MHA for capturing the relationship between acoustic information HH and textual information from the masked MHA. The residual connection [69] and layer normalization [70] are also employed inside each of the sub-decoder.

IV-B3 Transcriber Learning Objective

A combined CTC and sequence-to-sequence objective is conducted for the transcriber pre-training as shown below:

Transcriber=αCTC+(1α)S2S,CTC=LossCTC(Gctc,R),S2S=LossS2S(Gs2s,R)\begin{split}\mathcal{L}_{\text{Transcriber}}=\alpha\mathcal{L}^{\text{CTC}}+(1-\alpha)\mathcal{L}^{\text{S2S}},\\ \mathcal{L}^{\text{CTC}}=\text{Loss}_{\text{CTC}}(G_{ctc},R),\\ \mathcal{L}^{\text{S2S}}=\text{Loss}_{\text{S2S}}(G_{s2s},R)\end{split} (5)

where α[0,1]\alpha\in[0,1], RR is the ground-truth lyrical token sequence. The transcriber decoder are followed by a linear projection and softmax layers, that converts the decoder output OO into a posterior probability distribution of the predicted lyrical token sequence Gs2sG_{s2s}. The sequence-to-sequence (S2S) loss is the cross-entropy of ground-truth lyrical token sequence RR and its corresponding predicted lyrical token sequence Gs2sG_{s2s}. Also, a linear transform is applied on HH to obtain the token posterior distribution GctcG_{ctc}. CTC loss is computed between GctcG_{ctc} and RR [60].

IV-C Learning Objective

To leverage on the acoustic and linguistic knowledge from the pre-trained sRes extractor and transcriber models, PoLyScriber is initialized by the respective pre-trained models for training. PoLyScriber is trained to minimize S2S and CTC losses with an objective function PoLyScriber\mathcal{L}_{\text{PoLyScriber}},

PoLyScriber=αCTC+(1α)S2S\mathcal{L}_{\text{PoLyScriber}}=\alpha\mathcal{L}^{\text{CTC}}+(1-\alpha)\mathcal{L}^{\text{S2S}} (6)

During integrated fine-tuning, the parameters are differentiable and backpropogated all the way to the extractor towards the designed losses. For the variants of PoLyScriber, PoLyScriber-NoAug shares the same objective function with PoLyScriber, while PoLyScriber-L is holistically optimized by S2S, CTC and extraction losses with an objective function PoLyScriber-L\mathcal{L}_{\text{PoLyScriber-L}},

PoLyScriber-L=αCTC+(1α)S2S+Ext\mathcal{L}_{\text{PoLyScriber-L}}=\alpha\mathcal{L}^{\text{CTC}}+(1-\alpha)\mathcal{L}^{\text{S2S}}+\mathcal{L}^{\text{Ext}} (7)

At run-time inference, PoLyScriber and its variants directly converts the input polyphonic music to the output lyrical token sequence by passing through the globally trained network of the extractor and the transcriber.

V Experiments

V-A Dataset

Our experiments are conducted using four kinds of datasets - a polyphonic music dataset, a solo-singing dataset (clean vocals without background accompaniment), a music separation dataset, and a simulated polyphonic music dataset.

TABLE III: A description of solo-singing dataset.
Name # songs Lyrical lines Duration
Solo-train 4,324 81,092 149.1 hours
Solo-dev 66 482 0.7 hours
Solo-test 70 480 0.8 hours

V-A1 Polyphonic Music Dataset

We explore the task of lyrics transcription on polyphonic music. As shown in Table II, the polyphonic music training dataset, Poly-train, consists of the DALI-train [71] dataset and a NUS proprietary collection. The DALI-train dataset consists of 3,913 English polyphonic audio tracks111There are a total of 5,358 audio tracks in DALI, but we only have access to 3,913 English audio links.. The dataset is processed into 180,034 lyrics-transcribed audio lines with a total duration of 208.6 hours. The NUS collection dataset consists of 517 popular English songs. We obtain its line-level lyrics boundaries using the state-of-the-art audio-to-lyrics alignment system [18], leading to 26,462 lyrics-transcribed audio lines with a total duration of 27.0 hours.

TABLE IV: A summary of the baselines and the proposed integrated fine-tuning models and their lyrics transcription results (WER%) on both Poly-test and Poly-dev datasets.
Poly-dev Poly-test
Baseline Models Train data Extractor Transcriber Hansen Jamendo Mauch
pre-and-pre UMX [63] Solo-train pre-train UMX pre-train SoloM 97.21 97.87 98.00 97.59
pre-and-pre sRes [37] Solo-train pre-train sRes pre-train SoloM 67.77 66.76 62.48 58.28
pre-and-fine UMX Poly-train pre-train UMX fine-tune SoloM 63.06 64.35 58.24 78.31
pre-and-fine sRes Poly-train pre-train sRes fine-tune SoloM 44.33 42.03 48.00 34.89
Direct Modeling (DM) Poly-train fine-tune SoloM 44.95 40.15 44.77 38.13
DM-Aug Poly-train + DataAug fine-tune SoloM 46.70 35.61 44.75 36.43
Integrated Fine-tuning Models Train data Extractor Transcriber Poly-dev Hansen Jamendo Mauch
PoLyScriber-NoAug Poly-train fine-tune sRes fine-tune SoloM 39.47 34.91 41.38 31.34
PoLyScriber Poly-train + DataAug fine-tune sRes fine-tune SoloM 39.10 32.02 40.41 30.78
PoLyScriber-L Poly-train + DataAug fine-tune sRes fine-tune SoloM 40.25 32.58 40.13 32.77
  • DataAug denotes the simulated polyphonic music. sRes and UMX denote simplified Residual-UNet and the UMX extractors, respectively.

The Poly-dev, serving as the validation dataset, consists of the DALI-dev dataset of 100 songs from DALI dataset [18], and 70 songs from a NUS proprietary collection. We adopt three widely used test sets – Hansen[72], Jamendo[17], and Mauch[73] to form the Poly-test as shown in Table II. The test datasets are English polyphonic songs, that are manually segmented into line-level audio snippets of an average length of 8.13 seconds, each of which we will refer to as an audio line. In our experiments, we transcribe the lyrics of a song line-by-line following [42].

V-A2 Solo-singing Dataset

We study the use of pre-training on solo-singing for lyrics transcription. A curated version [41] of the English solo-singing dataset Sing! 300×30×2\textit{Sing! 300}\times\textit{30}\times\textit{2} 222The audio files can be accessed from https://ccrma.stanford.edu/damp/ is adopted, and detailed in Table III. A recent study [25] reports the state-of-the-art performance on this dataset, that serves as a good performance reference. As indicated in [41], the training set Solo-train consists of 4,324 songs with 81,092 audio lines. The validation set Solo-dev and the test set Solo-test contain 66 songs and 70 songs with 482 and 480 audio lines respectively. The lyrics of this database are also manually transcribed [41].

V-A3 Music Separation Database

We use a standard music separation database, MusDB18 [34], for the singing vocal extractor pre-training and evaluation in PoLyScriber and its variants. MusDB18 is a widely used database for singing voice separation and music source separation tasks, and it contains 150 full-length tracks with separate vocals and accompaniment where 86, 14 and 50 songs are designed for training, validation and testing, respectively. All songs are stereo with a sampling rate of 44.1 kHz, as described in [37].

V-A4 Simulated Polyphonic Music Dataset

For the purpose of data augmentation for training the transcriber, PoLyScriber and PoLyScriber-L, we create a simulated polyphonic music dataset. The simulated training set (4,324 songs) is generated by adding music tracks, selected at random from MusDB18[34], to every audio clip in Solo-train data at the time of training. Specifically, for each epoch of training, the solo-singing training set audio clips are mixed with random background accompaniment tracks at a wide range of signal-to-noise ratios (SNRs) sampled randomly between [-10dB, 20dB].

V-B Experimental Setup

We detail the network architectures of the extractors and the transcribers as well as the training and decoding parameter settings for the proposed PoLyScriber and its variants as well as reference baselines in the following. Reference baselines in Table IV consist of the two-step pipelines (pre-and-pre and pre-and-fine strategies) and DM as described in Section IV-B.

V-B1 Extractor

To understand the effects of different extractor architectures on lyrics transcription, we implement two systems - the main extractor is the simplified Residual-UNet (sRes) described in Fig. 1 and Section IV-A, which is one of the top performing systems for singing vocal extraction in Music Demixing Challenge 2021 [67], and another extractor for comparison is the Open-Unmix (UMX) [31] that was the best performing open-source music source separation system in the source separation challenge SiSEC 2018 [33]. We further detail the network architecture of the simplified Residual-UNet (sRes) and UMX below.

For Simplified Residual-UNet Vocal (sRes) Extractor, model compression is applied to reduce the model size and speed up the integrated fine-tuning process. The original Residual-UNet model [37] employs model architecture with a large parameter size of 102 M trainable parameters, which brings difficulties in model deployment due to the expensive and time-consuming computation. To reduce the model size, we apply model compression with several redundant layers removed while ensuring that the performance of vocal extraction does not get affected significantly. Compared to the original Residual UNet [37], in this work, we have removed two REBs, two ICBs and two RDBs where two RCBs are further removed in each REB and each RDB to establish the simplified Residual-UNet. The sRes extractor thus contains four REBs, two ICBs, and four RDBs with only 4.4 M trainable parameters.

Regarding to UMX Vocal Extractor, we utilize “UMX” model from Open-unmix serving as UMX vocal extractor [63] as a comparison system for the front end as it is a widely used open-source singing vocal extraction system. UMX is trained by parallel polyphonic mixture and clean singing vocal audio in MusDB18 dataset [34] using bidirectional Long short-term memory (BLSTM), and it learns to predict the magnitude spectrogram of singing vocal from the magnitude spectrogram of the corresponding mixture inputs (singing vocal+background music). The network consists of a 3-layer BLSTM where each layer has 512 nodes. UMX is optimized in the magnitude spectrogram level using mean square error, and the singing vocal prediction is obtained by applying a mask on the input polyphonic music.

We compare the two extractors briefly. The sRes and UMX extractors are both spectrogram-based approach built on top of the MusDB18 dataset [34] and produces the extracted vocal spectrogram, but the objective function of UMX is computed on spectrogram-level while that of sRes is on time-domain. The idea of obtaining extracted vocal spectrogram are also different. The sRes extractor estimates both phase and magnitude of the cIRM as well as incorporates the additional direct magnitude prediction to compensate the magnitude of cIRMs, while the UMX extractor does not predict the phases for extracted singing vocal that makes it suffer from incorrect phase reconstruction problem [37].

V-B2 Baseline Frameworks

Reference baselines are presented in Table IV and they include direct modeling (DM) as well as two-step pipeline approaches. Specifically for direct modeling method, DM and DM-Aug models are directly trained and validated on polyphonic music data, and they share the same architecture which only includes transformer-based transcriber that is pre-trained by solo-singing data.

There are two kinds of two-step pipeline strategies. First, in the pre-and-pre strategy [41, 49], the extractor and the transcriber are separately pre-trained. During inference, the trained extractor front end is used to get the extracted singing vocal that is then passed to the transcriber, that is trained on clean solo-singing vocals, for the decoding of the polyphonic music development and test sets. The second strategy, pre-and-fine, firstly employs pre-training on the front-end extractor for extracting vocals from polyphonic music, followed by fine-tuning solo-singing based transcriber using extracted vocal data [52]. The solo-singing based transcriber is pre-trained by solo-singing data. Polyphonic music data also goes through the extractor for testing the pre-and-fine model.

To be more specific, we use off-the-shelf SOTA systems for both the modules in our two-step baseline, where pre-and-pre and pre-and-fine approaches are experimented on two pre-trained extractors including simplified Residual-UNet and UMX. The pre-trained “UMX” [63] is utilized to extract singing vocals for the two-step pipelines pre-and-pre UMX and pre-and-fine UMX, and the pre-trained simplified Residual-UNet is employed to extract vocals for pre-and-pre sRes and pre-and-fine sRes models. We note that the audio inputs for the feature extraction block are real/simulated polyphonic music for DM or DM-aug models, and extracted singing vocal for pre-and-pre and pre-and-fine models.

V-B3 Integrated Fine-tuning Frameworks

Instead of directly training the transcriber on polyphonic music as in DM-aug, the proposed integrated fine-tuning approach alleviates the background music inference problem by integrating the extractor and the transcriber for training in a single network. Specifically, we conduct pre-training of the simplified Residual-UNet extractor first, and then globally train the pre-trained extractor and transcriber in a single network. The pre-training uses MusDB18 dataset [34] following the default parameter settings in [37], and the integrated fine-tuning of PoLyScriber and its variants (PoLyScriber-NoAug and PoLyScriber-L) uses the polyphonic music data as detailed in Table IV. Specifically, data augmentation is performed for PoLyScriber and PoLyScriber-L models but not for PoLyScriber-NoAug model. The feature extraction block inputs for PoLyScriber and its variants are intermediate extracted singing vocal audio from the extractors.

V-B4 Training and Inference Configurations

We use ESPnet [74] with PyTorch back end to build all transcribers, and the interpolation factor α\alpha between CTC loss and S2S loss is set to 0.3 for training. Other parameters of the transcriber follow the default setting in published LibriSpeech model (LS online)333See the pretrained librispeech model “pytorch large Transformer with specaug (4 GPUs) + Large LSTM LM” from the ESPNET githubhttps://github.com/espnet/espnet/blob/master/egs/librispeech/asr1/RESULTS.md., where attention dim is 512, the number of heads is 8 in MHA and FFN laryer dim is 2048. During integrated fine-tuning, the extractor input are upsampled to 44.1k sampling rate using the training data as indicated in Table IV. All models are trained using the Adam optimizer with a Noam learning rate decay as in [68], 25,000 warmup steps, 2,000,000 batch-bin, and 20 epochs [68]. We follow the default setting in ESPnet [74] to average the best 5 validated model checkpoints on the development set (Solo-dev for pre-and-pre models, librispeech dev dataset for LS model and Poly-dev for the rest models) to obtain the final model as in Table IV. We follow the common joint decoding approach [75, 60], which takes CTC prediction score into account during decoding by setting different CTC weight. During decoding process of different models, the default setting is used where the penalty, beam width and CTC decoding weight are set to 0.0, 10 and 0.3, respectively.

Since singing vocals extracted from polyphonic music are noisy version of clean solo-singing vocals, it is reasonable to initialize the polyphonic lyrics transcriber with a pre-trained solo-singing lyrics transcription model. We developed the pre-trained solo-singing transcription model (SoloM) by first training a speech recognition model (LS) on LibriSpeech dataset [76] following LS online model setting with 80-dim fbank features. We note that the pre-trained LS and SoloM models have the same transformer-based E2E architecture design as in Section IV-B.

VI Results and Discussion

We study the effects of pre-training, different extractor models, two-step strategies, integrated fine-tuning, data augmentation and the incorporation of the extraction loss on lyrics transcription of polyphonic music. In-depth spectrogram analysis, error analysis and music genre analysis are also performed to understand the behavior of lyrics transcription systems for different music genres. We further conduct ablation study on the proposed model to explore the contributions of each sub-component. Also, a comparative study of the proposed models with the existing approaches is presented.

We present a summary of the performance of the baseline approaches - the two-step models, and direct modeling approaches, as well as the proposed approach and its variants in Table IV. We also present different integrated fine-tuning mechanisms that include transcription-oriented training and inclusion of the extraction loss.

VI-A Evaluation Methods

We report the lyrics transcription performance in terms of word error rate (WER), which is the ratio of the total number of insertions, substitutions, and deletions with respect to the total number of words. Signal-to-Distortion ratio (SDR) is used for evaluating the singing voice extraction following [63, 37, 67], which is defined as:

SDR=10log10n𝐬(n)2+ϵn𝐬(n)𝐬^(n)2+ϵ,\centering SDR=10log_{10}\frac{\sum_{n}||\mathbf{s}(n)||^{2}+\epsilon}{\sum_{n}||\mathbf{s}(n)-\hat{\mathbf{s}}(n)||^{2}+\epsilon},\@add@centering (8)

where s(n)R\textbf{s}(n)\in R denotes the waveform of the ground truth solo-singing and s^(n)\hat{\textbf{s}}(n) is the waveform of the estimate for the extracted singing vocal with n being the (discrete) time index. We use a small constant ϵ=107\epsilon=10^{-7} in Eq. 8 to avoid divisions by zero. The higher the SDR score is, the better the output of the system is.

VI-B Performance of the Pre-trained Models

TABLE V: Comparison of speech recognition and lyrics recognition (WER%) performances between published online models and our models.
Pre-trained Speech Model LS-online LS
dev-clean 3.29 3.72
dev-other 8.49 9.91
test-clean 3.63 4.05
test-other 8.35 9.66
Pre-trained Solo-sing Model SOTA [25] SoloM
Solo-dev 18.74 16.33
Solo-test 14.79 16.30

We first present the pre-trained model performances of the extractor and the transcriber to prepare for constructing the proposed PoLyScriber.

As shown in Table V, the LS model is comparable in performance with the online published LS-online model. The pre-trained solo-singing model SoloM is initialized with the LS model, and then trained on solo-singing database as detailed in Table III. In Table V, we can see that the transformer-based SoloM model achieves competitive performance compared with the current published SOTA [25] reference model that is based on Kaldi speech recognition engine.

While developing the pre-trained model for the main extractor (we call it sRes in Table IV), we employ model compression on the original residual UNet [37] with 102 M trainable parameters and obtain a simplified Residual-UNet with 4.4 M trainable parameters as described in Section V-B1. We observe that the simplified Residual-UNet (sRes) achieves comparable performance of SDR 8.41 dB on widely used MusDB18 testset [34] in comparison with the original model [37] of SDR 8.90 dB.

In Table IV, we also compare the two-step approaches on two different extractors, a popular open-source UMX and recent high performing vocal extractor Residual-UNet, eg. pre-and-fine sRes vs. pre-and-fine UMX. Note that the simplified Residual-UNet outperforms UMX for the task of singing vocal extraction on MusDB18 testset with SDR of 8.41dB and 6.32dB, respectively. We observe a superior performance of the simplified Residual UNet extractor over the UMX extractor for both pre-and-pre and pre-and-fine approaches for lyrics transcription as well. This corroborates the intuition that an improved vocal extraction front end helps the downstream tasks, e.g. the lyrics transcription. To this regards, we perform the rest of the experiments using the simplified Residual-UNet extractor.

VI-C Performance of the Baseline Approaches

To explore the different strategies of incorporating singing vocal extractor for lyrics transcription, we report the lyrics transcription performances of two-step pipelines, i.e the pre-and-pre and pre-and-fine approaches.

In Table IV, we can see that the pre-and-fine approaches significantly outperform the corresponding pre-and-pre approaches on both the extractors (eg. pre-and-fine sRes with WER of 34.89% achieves relative 40.10% improvement from pre-and-pre sRes with WER of 58.25% on Mauch dataset). This consistent improvement from pre-and-pre approach to pre-and-fine approach indicates that pre-and-fine approach is more capable of addressing the mismatch problem that exists in the pre-and-pre approach between the training and testing for the transcriber.

We note that the DM approach outperforms pre-and-pre UMX, pre-and-pre sRes and pre-and-fine UMX systems across all datasets significantly, and performs better than pre-and-fine sRes for Hansen and Jamendo datasets. This suggests that the DM approach is as competitive as the two-steps pipelines, and the presence of music in DM actually contributes to compensate the distorted parts that is lack in two-step pipelines.

VI-D Performance of the Integrated Fine-tuning Approaches

Refer to caption
Figure 2: The spectrograms of intermediate extracted vocals from (a) PoLyScriber extractor and (b) PoLyScriber-L extractor; (c) the spectrogram of pre-and-fine sRes model input vocal; and (d) the spectrogram of DM input original polyphonic music. The same 2-second audio clip is used for plotting spectrograms as the example in Table VI with the transcription ”Just a taste of my bad side”.

VI-D1 Comparison of Integrated Fine-tuning and Two-Step Pipelines

We investigate the effect of the integrated fine-tuning by comparing the proposed PoLyScriber and its variants with the two-step approaches. In Table IV, we observe that PoLyScriber-NoAug framework consistently outperforms pre-and-fine sRes and pre-and-pre sRes models. This suggests that our integrated fine-tuning strategy of combining the training process of the extractor indeed yields better results in predicting lyrics compared to the two-step approaches, because both of these modules are optimized towards the common objective of lyrics transcription. Since the extractor in the two-step methods is not optimized towards the goal of lyrics transcription, the output from the extractor may not be a suitable input for the lyrics transcriber, leading to a sub-optimal solution. This is due to the fact that singing vocal extraction front ends are optimized to estimate the target singing vocal, which is different from optimizing for lyrics transcription. Therefore, the E2E PoLyScriber-NoAug, that optimizes the whole network towards lyrics transcription objective, performs better, and is able to address the mismatch problem between the front end and back end in the two-step approaches.

Furthermore, the superiority and effectiveness of the PoLyScriber-NoAug model over pre-and-fine sRes model suggests that the integrated fine-tuning approach is capable of performing effective transcription using only real-world polyphonic music data independently, without needing any parallel singing vocal and polyphonic music data generated from data augmentation.

TABLE VI: Examples of transcribed lyrics with five representative models. # C I D S denote the number of correct words (C), substitutions (S), insertions (I) and deletions (D).
reference (ref): get a taste of my bad side just a taste of my bad side just a taste of my bad side
# C I D S WER (%)
pre-and-fine sRes
hypothesis (hyp): get your text up and touch the taste of
3 0 12 6 85.71
DM
hyp: get a taste of my bad side Im just a taste of my bad side
11 0 6 4 47.62
PoLyScriber-NoAug
hyp: get a taste of my bad time just a taste of my bad side
12 0 7 2 42.86
PoLyScriber
hyp: get a taste of my bad times into the taste of my body just a taste of my
14 0 3 4 33.33
PoLyScriber-L
hyp: get a taste of my outside and just a taste of my bad time just a taste of my
16 0 2 3 23.81

VI-D2 Comparison of Integrated Fine-tuning and Direct Modeling

To study the importance of tackling music interference problem for lyrics transcription, we compare direct modeling (DM) approach with integrated fine-tuning approach. DM method directly trains the transcriber on polyphonic music data without considering the background music as an interference. In Table IV, we observed that PoLyScriber-NoAug outperforms the DM method across all the datasets, which suggests that lyrics transcription can benefit from the integrated fine-tuning process with the extractor to handle the music interference problem.

It is interesting to note that the DM method has a comparable, sometimes even better, performance than the two-step methods. This has also been observed previously in [51, 18]. This may be an indication that background music may not be as harmful as the vocal artifacts introduced by the vocal extractor in lyrics transcription [18, 42]. This observation also indicates that the integrated fine-tuning would finally optimize the extractor into a state where its output will be somewhere in-between absolute background music suppression and complete background music presence, so as to gain from the benefits of both.

TABLE VII: Analysis of lyrics recognition performance in terms of errors (substitutions (S), insertions (I) and deletions (D).) for Poly-test. These are the number of words in error and the percentage of each error with respect to the total number of words.
pre-and-fine sRes DM PoLyScriber-NoAug PoLyScriber PoLyScriber-L
Poly-test # I # D # S # I # D # D # I # D # S # I # D # S # I # D # S
All 22 248 229 24 188 253 24 182 225 35 115 257 36 104 247
All / # total words (%) 0.16 1.87 1.73 0.18 1.42 1.91 0.19 1.37 1.70 0.26 0.87 1.94 0.28 0.78 1.87

VI-D3 Data Augmentation

Data augmentation is employed by randomly adding music to solo-singing to create simulated polyphonic music data while training the transcriber. We use this data augmentation method in direct modeling and integrated fine-tuning approaches. In Table IV, we can see that DM-Aug outperforms DM for all the three testsets, and PoLyScriber outperforms PoLyScriber-NoAug. This indicates that the proposed singing data augmentation is a simple yet beneficial solution to the problem of lack of diversity in training data, thereby improving model generalization.

VI-D4 The Inclusion of the Extraction Loss

Having access to the parallel solo-singing and simulated polyphonic music, we are able to incorporate the extraction loss for the integrated fine-tuning framework. We investigate if the additional extraction loss is helpful to the integrated fine-tuning. In Table IV, we can see that the incorporation of the extraction loss in the PoLyScriber-L model shows a slight performance degradation on hansen and mauch datasets. Similar observations have been made in ASR studies where ASR-oriented optimization is effective for multi-talker speech recognition but an additional separation loss would not bring improvement to the recognition performance [77, 78, 79].

To analyze the relationship between the transcriber and the extractor in the integrated fine-tuning approaches, we present the extracted singing vocals from different systems and the corresponding original polyphonic music, along with their predicted transcriptions at the link 444https://xiaoxue1117.github.io/PaperSample/. Qualitatively, when we listened to the outputs of the extractor after integrated fine-tuning, we found that some background music/noise got added back into the audio and distorted vocal parts in two-step approaches were recovered in integrated fine-tuning approach. Some interesting observations were that the lyrics transcription improved, compared to two-step methods, especially in the chorus sections of the song where multiple voices were present simultaneously. The integrated fine-tuning seemed to have drowned out the other vocals present in the background by adding back some music/noises. This gives evidence that the proposed integrated network adds in some background music/noise and settles somewhere between completely suppressed background music (as in two-step approaches) and completely present background music (as in DM approaches), yielding better results than both. Similar observation are also presented through spectrogram visualization in the following Section VI-E.

TABLE VIII: The genre distribution for polyphonic music Dataset, genre-specific average word duration per utterance and lyrics transcription performance (WER%) on polyphonic music test sets.
Statistics Metal Pop Hip hop
# songs Hansen 3 6 1
# songs Jamendo 7 9 4
# songs Mauch 8 12 0
Percentage in Poly-train 35% 59% 6%
Percentage in Poly-dev 48% 49% 3%
Percentage in Poly-test 34% 56% 10%
Word duration in Poly-train (seconds) 0.96 0.90 0.65
Word duration in Poly-dev (seconds) 0.76 0.69 0.58
Word duration in Poly-test (seconds) 0.78 0.74 0.44
Model Metal Pop Hip hop
pre-and-fine UMX 75.55 64.43 69.44
pre-and-fine sRes 49.83 35.31 60.85
DM 52.17 36.85 52.69
DM-Aug 51.73 35.55 50.39
PoLyScriber-NoAug-Pop 46.47 34.53 54.04
PoLyScriber-NoAug-Hiphop 63.85 47.18 59.16
PoLyScriber-NoAug-Metal 47.56 35.68 58.46
PoLyScriber-NoAug 43.70 32.40 51.54
PoLyScriber 46.16 31.20 46.55
PoLyScriber-L 46.79 32.07 46.88

VI-E Error Analysis and Spectrogram Visualization

We show an example in Table VI and conduct an error analysis on the Poly-test in Table VII. We also visualize the magnitude spectrogram of a 2-second snippet for the example in Fig. 2. It is noted that the number of deletions in the proposed PoLyScriber and its variants is substantially reduced from that of the pre-and-fine and DM approaches. Furthermore, PoLyScriber outperforms PoLyScriber-NoAug, which suggests the integrated fine-tuning helps to recover words that are deleted by the traditional methods. This again confirms the effectiveness of the integrated fine-tuning and data augmentation.

We can observe from Fig. 2 that the DM input spectrogram Fig. 2 (d) contains loud music accompaniments and pre-and-fine sRes input spectrogram Fig. 2 (c) suffers from the vocal distortion especially for the duration of 0s-0.75s and 1.5s-2s, which has been alleviated by the proposed integrated fine-tuning approach as in Fig. 2 (a) (b). By comparing Fig. 2 (d) with Fig. 2 (a) and Fig. 2 (b), we observe that the background accompaniments are removed from the spectrograms by the PoLyScriber and its variants, leading to improved lyrics transcription. The audio samples are available online 555https://xiaoxue1117.github.io/PaperSample/.

VI-F Music Genre Analysis

We analyse the performance of different music genres per song on the polyphonic test sets – Hansen, Jamendo and Mauch. The music genre distribution and average lyrical word duration per utterance in poly-train, poly-dev and poly-test is summarized in Table VIII, and the test datasets consist of three broad genre categories – pop, hip hop and metal, as given in [18]. We report the lyrics transcription performance by music genre for our best performing models in Table VIII.

We observe that the proposed PoLyScriber and its variants consistently outperform the DM-based approaches and two-step approaches across all the genres, which further verifies the effectiveness of the integrated fine-tuning technique over the conventional approaches. PoLyScriber-NoAug model without data augmentation performs the best for the metal songs, which suggests that the SNR-based diversity generated in the training data through our augmentation method does not appropriately generalize the model for metal songs. Metal songs often contain loud distorted guitar sounds which is not that common in pop and hip-hop songs. The kinds of instruments used in metal songs differ from pop or hip-hop songs, and the music tracks used for augmentation were mostly from pop genre, which may have been the reason of this drop in performance for metal songs. For hip-hop songs, data augmentation involved models including PoLyScriber and PoLyScriber-L perform better than other models without data augmentation. This suggests that data augmentation provides more diversity and robustness that helps in lyrics transcription of hip-hop songs.

TABLE IX: The genre distribution for MusDB18 test set and vocal extractor performance (SDR dB) on MusDB18 test set.
Model Metal Pop Hip hop
PoLyScriber-NoAug 1.383 1.565 1.419
PoLyScriber-NoAug-Pop 1.502 1.613 1.073
PoLyScriber-NoAug-Hiphop 5.320 4.231 6.438
PoLyScriber-NoAug-Metal 2.783 3.145 3.448
PoLyScriber -0.983 -0.876 -1.783
PoLyScriber-L 0.547 0.557 0.106
Statistics Metal Pop Hip hop
# songs in MusDB18 testset 33 15 2
TABLE X: Ablation study of PoLyScriber-L on lyrics transcription performance (WER%) of polyphonic music.
Ablation Model Extractor Transcriber Poly-dev Hansen Jamendo Mauch
PoLyScriber-L fine-tune fine-tune 40.25 32.58 40.13 32.77
-ExtLoss (PoLyScriber) fine-tune fine-tune 39.10 32.02 40.41 30.78
-ExtLoss, DataAug (PoLyScriber-NoAug) fine-tune fine-tune 39.47 34.91 41.38 31.34
-ExtLoss, DataAug, Pre-train extractor (model A) scratch fine-tune 42.75 39.38 44.06 44.27
-ExtLoss, DataAug, Pre-train extractor-transcriber (model B) scratch scratch 55.80 62.82 65.49 67.37
-ExtLoss, DataAug, fine-tune extractor (pre-and-fine sRes) fix fine-tune 44.33 42.03 48.00 34.89
-ExtLoss, DataAug, Extractor (DM) remove fine-tune 44.95 40.15 44.77 38.13
-ExtLoss, DataAug, Extractor, fine-tune transcriber (DM-Scratch) remove scratch 53.52 57.66 65.68 48.17
  • 1

    Notes: - denotes the removal of certain components from the PoLyScriber-L model.

Furthermore, we observe from Table VIII that metal songs have the longest lyrical word duration, and lyrical word duration for hip-hop songs is significantly lower than that of metal and pop songs across all datasets. This indicates that the hip-hop songs show higher syllable rates, and genre affects lyrical words in terms of speaking rate. As described in [80, 18], metal and hip-hop songs show lower lyrics intelligence compared with pop songs. Specifically, metal songs possess louder background music than pop songs and ”Death Metal” songs shows zero lyrics intelligibility score [80]. Hip-hop songs that always consist of rap with many words have a higher syllable rate and rapid vocalization than pop songs, thereby receiving lower lyrics intelligibility scores [80, 18]. Moreover, the percentage of available training data for pop songs is much more than metal and hip-hop songs. This explains why hip-hop and metal songs have a higher word error rate than pop songs across all models in Table VIII.

To better understand model generalization ability across genres for both extractor and transcriber, genre-specific integrated fine-tuning (NoAug) models are trained using genre-specific data from Poly-train while tested on all the genres. For example, the PoLyScriber-NoAug-Pop model is trained on pop songs from Poly-train and tested on all the genres from Poly-test. We present extractor performance on MusDB18 test set together with its music genre distribution in Table IX and transcriber performance in Table VIII. We observe that the model with the extraction loss (PoLyScriber-L) generally performs better than the model without the extraction loss (PoLyScriber) on the singing vocal extraction task. This indicates that the inclusion of the extraction loss to integrated fine-tuning is beneficial to singing vocal extraction task. Table VI shows that PoLyScriber-NoAug-Hiphop for extraction purpose generalizes well on metal and pop songs but it performs not well on transcription performance in Table VIII, and other genre-specific models also show good generalization ability on other genres that are not covered in the training set. This indicates that there is a tradeoff between extractor and transcriber for PoLyScriber-NoAug-Hiphop, and it is not well-trained towards lyrics transcription purpose with less training data while it remains the pre-trained vocal extractor performance. We can observe from Table VIII that genre-specific models show worse generalization ability than the proposed model and its variants in terms of lyrics transcription performance, while genre-specific models perform better than the proposed models for vocal extraction in Table IX. This indicates that there is a trade-off between the extractor and the transcriber performances while models converge, wherein transcription performance improves when the extractor performs worse.

TABLE XI: Comparison between the proposed solutions and other existing competitive solutions on whole song test to lyrics transcription (WER%) of polyphonic music.
Kaldi-based Models Hansen Jamendo Mauch Dali-test
RB1 [50] 83.43 86.70 84.98 -
DDA2 [81] 74.81 72.15 75.39 -
DDA3 [81] 77.36 73.09 80.66 -
CG [18] - 59.60 44.00 -
GGL2 [82] 48.11 61.22 45.35 -
GGL1 [82] 45.87 56.76 43.76 -
MSTRE-Net [53] 36.78 34.94 37.33 42.11
End-to-End Models Hansen Jamendo Mauch Dali-test
DS [17] - 77.80 70.90 -
E2E Transformer [42] 40.02 45.19 38.96 46.56
Multi-transcriber [42] 36.85 44.12 33.69 40.20
PoLyScriber-NoAug (ours) 35.77 41.10 33.39 40.45
PoLyScriber (ours) 33.57 41.32 32.83 36.52
PoLyScriber-L (ours) 33.46 41.00 34.87 37.01

VI-G Ablation Study

To study where the contributions come from and verify the effectiveness of the integrated fine-tuning, we conduct an ablation study on the integrated fine-tuning model with extraction loss (PoLyScriber-L) in Table X. The removal of extraction loss brings slight improvement for lyrics transcription performance, which further verifies the advantage of transcription-oriented optimization for integrated fine-tuning. Moreover, we wonder if we can train a singing vocal extractor without parallel vocal and mixture data via integrated fine-tuning system. And can we even train the system from scratch?

A clear benefit of fine-tuning the extractor can be noticed while removing the extractor fine-tuning of PoLyScriber-NoAug. Concerning the lack of the extractor pre-training, model A still outperforms pre-and-fine sRes and DM for two testsets. To this regards, the performance of model A is acceptable and it is possible to train a singing vocal extractor without parallel vocal and mixture data inside the integrated fine-tuning framework. We can further observe that the model B suffers from a big performance drop from model A, which shows it is essential to pre-train the transcriber.

TABLE XII: Comparison of lyrics transcription (WER%) of polyphonic music on line-level testsets
Line-level test Hansen Jamendo Mauch
E2E Transformer [42] 39.87 44.26 36.80
Multi-transcriber [42] 36.34 43.44 31.82
PoLyScriber-NoAug (ours) 34.91 41.38 31.34
PoLyScriber (ours) 32.02 40.41 30.78
PoLyScriber-L (ours) 32.58 40.13 32.77

To compare the two-step pipelines with integrated fine-tuning, we can find that the lyrics transcription performance drops significantly from integrated fine-tuning approach (PoLyScriber-NoAug) to pre-and-fine sRes and DM. This indicates that integrated fine-tuning is critical to improve lyrics transcription performance. We also notice that the pre-training for the transcriber is also important while comparing DM with DM-Scratch.

VI-H Comparison with the State-of-the-Art

We compare PoLyScriber and its variants with the existing approaches for lyrics transcription on the polyphonic music testsets in Table XI. Specifically, we would like to compare PoLyScriber and its variants with the SOTA reference models [17, 18, 50, 81, 82, 53]. Stoller et al.’s [17] system is based on E2E Wave-U-Net framework and  [42] is built on an E2E transformer with multi-transcriber solutions. The rest of the systems [18, 50, 81, 82, 53] are all based on the traditional Kaldi-based ASR approach.

In Table XI, we first report the lyrics transcription performance of the existing systems on the same test sets for whole songs evaluation. We decode short nonoverlapping segments of songs in Hansen, Jamendo and Mauch using our proposed models and combine the transcriptions of these segments to report WER results for the complete songs as in [42]. We also test on a larger database DALI-test [53] with 240 whole-song polyphonic recordings processed as in [42]. We observe that the proposed PoLyScriber outperforms all previous E2E and Kaldi-based approaches for Hansen, Mauch and the large set DALI-test, which shows the general superiority of the integrated fine-tuning model over the conventional pipelines. To avoid segmentation problems, we further consider the performance of E2E models on line-level test sets for comparison as in [42] in Table XII, which shows that the proposed PoLyScriber and its variants outperform the E2E models [42], thereby achieving better performance among all the existing models. We also note that the PoLyScriber-NoAug without data augmentation performs better than E2E models, which shows the proposed integrated fine-tuning is capable of achieving good lyrics transcription performance without the need of data augmentation where the parallel solo-singing and polyphonic music is created.

VII Conclusion

PoLyScriber serves as an important step into the exploration of lyrics transcription for polyphonic music via an E2E integrated fine-tuning setting. We propose and idea to globally train singing voice extraction with lyrics transcriber towards the lyrics transcription objective for polyphonic music for the first time. We advocate the novel PoLyScriber framework with data augmentation for lyrics transcription of polyphonic music, that is proven to be effective. The proposed data augmentation paradigm enables PoLyScriber to leverage diverse polyphonic pattern and music knowledge from simulated polyphonic music for hip-hop songs. We have shown that the proposed PoLyScriber and its variants outperform baseline frameworks for lyrics transcription through a comprehensive set of experiments on publicly available datasets.

References

  • [1] A. J. Good, F. A. Russo, and J. Sullivan, “The efficacy of singing in foreign-language learning,” Psychology of Music, vol. 43, no. 5, pp. 627–640, 2015.
  • [2] Hang Zhao, Chuang Gan, Andrew Rouditchenko, Carl Vondrick, Josh McDermott, and Antonio Torralba, “The sound of pixels,” in Proceedings of the European conference on computer vision, 2018, pp. 570–586.
  • [3] Hang Zhao, Chuang Gan, Wei-Chiu Ma, and Antonio Torralba, “The sound of motions,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, 2019, pp. 1735–1744.
  • [4] Chuang Gan, Deng Huang, Hang Zhao, Joshua B Tenenbaum, and Antonio Torralba, “Music gesture for visual sound separation,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2020, pp. 10478–10487.
  • [5] Chuang Gan, Deng Huang, Peihao Chen, Joshua B Tenenbaum, and Antonio Torralba, “Foley music: Learning to generate music from videos,” in European Conference on Computer Vision. Springer, 2020, pp. 758–775.
  • [6] Zui Chen, Yansen Jing, Shengcheng Yuan, Yifei Xu, Jian Wu, and Hang Zhao, “Sound2synth: Interpreting sound via fm synthesizer parameters estimation,” Proc of the Thirty-First International Joint Conference on Artificial Intelligence AI and Arts, pp. 4921–4928, 2022.
  • [7] LAUREN B COLLISTER and DAVID HURON, “Comparison of word intelligibility in spoken and sung phrases,” Empirical Musicology Review, vol. 3, no. 3, 2008.
  • [8] G. Dzhambazov, Knowledge-based Probabilistic Modeling for Tracking Lyrics in Music Audio Signals, Ph.D. thesis, Universitat Pompeu Fabra, 2017.
  • [9] T. Hosoya, M. Suzuki, A. Ito, S. Makino, L. A. Smith, D. Bainbridge, and I. H. Witten, “Lyrics recognition from a singing voice based on finite state automaton for music information retrieval.,” in ISMIR, 2005, pp. 532–535.
  • [10] Philip A Fine and Jane Ginsborg, “Making myself understood: perceived factors affecting the intelligibility of sung text,” Frontiers in psychology, vol. 5, pp. 809, 2014.
  • [11] Hiromasa Fujihara and Masataka Goto, “Lyrics-to-audio alignment and its application,” in Dagstuhl Follow-Ups. Schloss Dagstuhl-Leibniz-Zentrum fuer Informatik, 2012, vol. 3.
  • [12] Xiaoxue Gao, Xiaohai Tian, Yi Zhou, Rohan Kumar Das, and Haizhou Li, “Personalized singing voice generation using wavernn.,” in Odyssey, 2020, pp. 252–258.
  • [13] Xiaoxue Gao, Xiaohai Tian, Rohan Kumar Das, Yi Zhou, and Haizhou Li, “Speaker-independent spectral mapping for speech-to-singing conversion,” in IEEE APSIPA ASC, 2010, pp. 159–164.
  • [14] Karthika Vijayan, Xiaoxue Gao, and Haizhou Li, “Analysis of speech and singing signals for temporal alignment,” in IEEE APSIPA ASC, 2018, pp. 1893–1898.
  • [15] Bidisha Sharma and Ye Wang, “Automatic evaluation of song intelligibility using singing adapted stoi and vocal-specific features,” IEEE/ACM Transactions on Audio, Speech, and Language Processing, vol. 28, pp. 319–331, 2020.
  • [16] Karim M Ibrahim, David Grunberg, Kat Agres, Chitralekha Gupta, and Ye Wang, “Intelligibility of sung lyrics: A pilot study.,” in ISMIR, 2017, pp. 686–693.
  • [17] D. Stoller, S. Durand, and S. Ewert, “End-to-end lyrics alignment for polyphonic music using an audio-to-character recognition model,” in IEEE ICASSP, 2019, pp. 181–185.
  • [18] C. Gupta, E. Yılmaz, and H. Li, “Automatic lyrics alignment and transcription in polyphonic music: Does background music help?,” in IEEE ICASSP, 2020, pp. 496–500.
  • [19] C.⁢ Gupta, B.⁢ Sharma, H. Li, and Y. Wang, “Automatic lyrics-to-audio alignment on polyphonic music using singing-adapted acoustic models,” in IEEE ICASSP, 2019, pp. 396–400.
  • [20] A. Mesaros and T. Virtanen, “Automatic recognition of lyrics in singing,” EURASIP Journal on Audio, Speech, and Music Processing, vol. 2010, no. 1, pp. 546047, 2010.
  • [21] G. B. Dzhambazov and X. Serra, “Modeling of phoneme durations for alignment between polyphonic audio and lyrics,” in Sound and Music Computing Conference, 2015, pp. 281–286.
  • [22] H. Fujihara, M. Goto, J. Ogata, and H. G. Okuno, “Lyricsynchronizer: Automatic synchronization system between musical audio signals and lyrics,” IEEE Journal of Selected Topics in Signal Processing, vol. 5, no. 6, pp. 1252–1261, 2011.
  • [23] Xiaoxue Gao, Chitralekha Gupta, and Haizhou Li, “Genre-conditioned acoustic models for automatic lyrics transcription of polyphonic music,” in IEEE ICASSP, 2022, pp. 791–795.
  • [24] Emir Demirel, Sven Ahlbäck, and Simon Dixon, “Low resource audio-to-lyrics alignment from polyphonic music recordings,” in IEEE ICASSP, 2021, pp. 586–590.
  • [25] Emir Demirel, Sven Ahlbäck, and Simon Dixon, “Automatic lyrics transcription using dilated convolutional neural networks with self-attention,” in IEEE IJCNN, 2020, pp. 1–8.
  • [26] Chitralekha Gupta, Haizhou Li, and Masataka Goto, “Deep learning approaches in topics of singing information processing,” IEEE/ACM Transactions on Audio, Speech, and Language Processing, vol. 30, pp. 2422–2451, 2022.
  • [27] Daniel Stoller, Sebastian Ewert, and Simon Dixon, “Wave-u-net: A multi-scale neural network for end-to-end audio source separation,” in ISMIR, 2018, pp. 334–340.
  • [28] Francesc Lluís, Jordi Pons, and Xavier Serra, “End-to-end music source separation: Is it possible in the waveform domain?,” in Interspeech, 2019, pp. 4619–4623.
  • [29] David Samuel, Aditya Ganeshan, and Jason Naradowsky, “Meta-learning extractors for music source separation,” in IEEE ICASSP, 2020, pp. 816–820.
  • [30] Alexandre Défossez, Nicolas Usunier, Léon Bottou, and Francis Bach, “Music source separation in the waveform domain,” arXiv preprint arXiv:1911.13254, 2019.
  • [31] Fabian-Robert Stöter, Stefan Uhlich, Antoine Liutkus, and Yuki Mitsufuji, “Open-unmix-a reference implementation for music source separation,” Journal of Open Source Software, vol. 4, no. 41, pp. 1667, 2019.
  • [32] Naoya Takahashi, Nabarun Goswami, and Yuki Mitsufuji, “Mmdenselstm: An efficient combination of convolutional and recurrent neural networks for audio source separation,” in 16th International Workshop on Acoustic Signal Enhancement. IEEE, 2018, pp. 106–110.
  • [33] Fabian-Robert Stöter, Antoine Liutkus, and Nobutaka Ito, “The 2018 signal separation evaluation campaign,” in International Conference on Latent Variable Analysis and Signal Separation. Springer, 2018, pp. 293–305.
  • [34] Zafar Rafii, Antoine Liutkus, Fabian-Robert Stöter, Stylianos Ioannis Mimilakis, and Rachel Bittner, “The MUSDB18 corpus for music separation,” Access on: Dec, 2017. [ONLINE] Available: https://doi.org/10.5281/zenodo.1117372. DOI: 10.5281/zenodo.1117372.
  • [35] Romain Hennequin, Anis Khlif, Felix Voituret, and Manuel Moussallam, “Spleeter: a fast and efficient music source separation tool with pre-trained models,” Journal of Open Source Software, vol. 5, no. 50, pp. 2154, 2020.
  • [36] Naoya Takahashi and Yuki Mitsufuji, “D3net: Densely connected multidilated densenet for music source separation,” arXiv preprint arXiv:2010.01733, 2020.
  • [37] Qiuqiang Kong, Yin Cao, Haohe Liu, Keunwoo Choi, and Yuxuan Wang, “Decoupling magnitude and phase estimation with deep resunet for music source separation,” in ISMIR, 2021, pp. 342–349.
  • [38] Laure Prétet, Romain Hennequin, Jimena Royo-Letelier, and Andrea Vaglio, “Singing voice separation: A study on training data,” in IEEE ICASSP, 2019, pp. 506–510.
  • [39] Xiaoxue Gao, Xianghu Yue, and Haizhou Li, “Self-transriber: Few-shot lyrics transcription with self-training,” arXiv preprint arXiv:2211.10152, 2022.
  • [40] A. Povey, D.and Ghoshal, G. Boulianne, L. Burget, O. Glembek, N. Goel, M. Hannemann, P. Motlicek, Y. Qian, P. Schwarz, J. Silovsky, G. Stemmer, and K. Vesely, “The Kaldi speech recognition toolkit,” in ASRU, 2011.
  • [41] Gerardo Roa Dabike and Jon Barker, “Automatic lyric transcription from karaoke vocal tracks: Resources and a baseline system.,” in Interspeech, 2019, pp. 579–583.
  • [42] Xiaoxue Gao, Chitralekha Gupta, and Haizhou Li, “Automatic lyrics transcription of polyphonic music with lyrics-chord multi-task learning,” IEEE/ACM Transactions on Audio, Speech, and Language Processing, vol. 30, pp. 2280–2294, 2022.
  • [43] Annamaria Mesaros and Tuomas Virtanen, “Adaptation of a speech recognizer for singing voice,” in 2009 17th European Signal Processing Conference. IEEE, 2009, pp. 1779–1783.
  • [44] Annamaria Mesaros, “Singing voice identification and lyrics transcription for music information retrieval invited paper,” in IEEE Conference on Speech Technology and Human-Computer Dialogue, 2013, pp. 1–10.
  • [45] Dairoku Kawai, Kazumasa Yamamoto, and Seiichi Nakagawa, “Speech analysis of sung-speech and lyric recognition in monophonic singing,” in IEEE ICASSP, 2016, pp. 271–275.
  • [46] A. M. Kruspe, “Bootstrapping a system for phoneme recognition and keyword spotting in unaccompanied singing.,” in ISMIR, 2016, pp. 358–364.
  • [47] Chitralekha Gupta, Haizhou Li, and Ye Wang, “Automatic pronunciation evaluation of singing.,” in Interspeech, 2018, pp. 1507–1511.
  • [48] Che-Ping Tsai, Yi-Lin Tuan, and Lin-shan Lee, “Transcribing lyrics from commercial song audio: the first step towards singing content processing,” in IEEE ICASSP, 2018, pp. 5749–5753.
  • [49] Chitralekha Gupta, Rong Tong, Haizhou Li, and Ye Wang, “Semi-supervised lyrics and solo-singing alignment.,” in ISMIR, 2018, pp. 600–607.
  • [50] Gerardo Roa Dabike and Jon Barker, “The sheffield university system for the MIREX 2020: Lyrics transcription task,” in MIREX, 2020.
  • [51] Chitralekha Gupta, Bidisha Sharma, Haizhou Li, and Ye Wang, “Lyrics-to-audio alignment using singing-adapted acoustic models and non-vocal suppression,” Music Inf. Retrieval Eval. eXchange Audio-Lyrics Alignment Challenge, 2022, Accessed on: Jul. 19, 2022. [Online]. Available: https://www. music-ir.org/mirex/abstracts/2018/GSLW1.pdf.
  • [52] C. Gupta, E. Yılmaz, and H. Li, “Acoustic modeling for automatic lyrics-to-audio alignment,” in Interspeech, 2019.
  • [53] Emir Demirel, Sven Ahlbäck, and Simon Dixon, “Mstre-net: Multistreaming acoustic modeling for automatic lyrics transcription,” in ISMIR, 2021, pp. 151–158.
  • [54] Xiaoxue Gao, Chitralekha Gupta, and Haizhou Li, “Music-robust automatic lyrics transcription of polyphonic music,” in Sound and Music Computing, 2022, pp. 325–332.
  • [55] Alex Graves and Navdeep Jaitly, “Towards end-to-end speech recognition with recurrent neural networks,” in International conference on machine learning. PMLR, 2014, pp. 1764–1772.
  • [56] Yajie Miao, Mohammad Gowayyed, and Florian Metze, “Eesen: End-to-end speech recognition using deep rnn models and wfst-based decoding,” in IEEE ASRU, 2015, pp. 167–174.
  • [57] Jan Chorowski, Dzmitry Bahdanau, Kyunghyun Cho, and Yoshua Bengio, “End-to-end continuous speech recognition using attention-based recurrent nn: First results,” in NIPS Workshop on Deep Learning, December, 2014.
  • [58] William Chan, Navdeep Jaitly, Quoc Le, and Oriol Vinyals, “Listen, attend and spell: A neural network for large vocabulary conversational speech recognition,” in IEEE ICASSP, 2016, pp. 4960–4964.
  • [59] Suyoun Kim, Takaaki Hori, and Shinji Watanabe, “Joint ctc-attention based end-to-end speech recognition using multi-task learning,” in IEEE ICASSP, 2017, pp. 4835–4839.
  • [60] Shigeki Karita, Nelson Enrique Yalta Soplin, Shinji Watanabe, Marc Delcroix, Atsunori Ogawa, and Tomohiro Nakatani, “Improving transformer-based end-to-end speech recognition with connectionist temporal classification and language model integration,” in Interspeech, 2019, pp. 1408–1412.
  • [61] A. Jansson, E. J. Humphrey, N. Montecchio, R. M. Bittner, A. Kumar, and T. Weyde, “Singing voice separation with deep U-Net convolutional networks,” in ISMIR, 2017, pp. 745–751.
  • [62] P. Chandna, M. Miron, J. Janer, and E. Gómez, “Monoaural audio source separation using deep convolutional neural networks,” in International Conference on Latent Variable Analysis and Signal Separation. Springer, 2017, pp. 258–266.
  • [63] Fabian-Robert Stöter, Stefan Uhlich, Antoine Liutkus, and Yuki Mitsufuji, “Open-unmix-a reference implementation for music source separation,” Journal of Open Source Software, vol. 4, no. 41, pp. 1667, 2019.
  • [64] Hyeong-Seok Choi, Jang-Hyun Kim, Jaesung Huh, Adrian Kim, Jung-Woo Ha, and Kyogu Lee, “Phase-aware speech enhancement with deep complex U-Net,” in International Conference on Learning Representations, 2018, https://openreview.net/forum?id=SkeRTsAcYm.
  • [65] DeLiang Wang and Jitong Chen, “Supervised speech separation based on deep learning: An overview,” IEEE/ACM Transactions on Audio, Speech, and Language Processing, vol. 26, no. 10, pp. 1702–1726, 2018.
  • [66] Ke Tan and DeLiang Wang, “Complex spectral mapping with a convolutional recurrent network for monaural speech enhancement,” in IEEE ICASSP, 2019, pp. 6865–6869.
  • [67] Yuki Mitsufuji, Giorgio Fabbro, Stefan Uhlich, Fabian-Robert Stöter, Alexandre Défossez, Minseok Kim, Woosung Choi, Chin-Yun Yu, and Kin-Wai Cheuk, “Music demixing challenge 2021,” Frontiers in Signal Processing, vol. 1, pp. 18, 2022.
  • [68] Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N. Gomez, Lukasz Kaiser, and Illia Polosukhin, “Attention is all you need,” in Neural Information Processing Systems, 2017, pp. 5998–6008.
  • [69] Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun, “Deep residual learning for image recognition,” in IEEE conference on computer vision and pattern recognition, 2016, pp. 770–778.
  • [70] Jimmy Lei Ba, Jamie Ryan Kiros, and Geoffrey E Hinton, “Layer normalization,” arXiv preprint arXiv:1607.06450, 2016.
  • [71] G. Meseguer-Brocal, A. Cohen-Hadria, and G. Peeters, “Dali: A large dataset of synchronized audio, lyrics and notes, automatically created using teacher-student machine learning paradigm,” in ISMIR, 2018, pp. 431–437.
  • [72] J. K. Hansen, “Recognition of phonemes in a-cappella recordings using temporal patterns and mel frequency cepstral coefficients,” in Sound and Music Computing, 2012, pp. 494–499.
  • [73] M. Mauch, H. Fujihara, and M. Goto, “Lyrics-to-audio alignment and phrase-level segmentation using incomplete internet-style chord annotations,” in Sound and Music Computing, 2010, pp. 9–16.
  • [74] S. Watanabe, T. Hori, S. Karita, T. Hayashi, J. Nishitoba, Y. Unno, N. Enrique Yalta Soplin, J. Heymann, M. Wiesner, N. Chen, A. Renduchintala, and T. Ochiai, “Espnet: End-to-end speech processing toolkit,” in Interspeech, 2018, pp. 2207–2211.
  • [75] Takaaki Hori, Jaejin Cho, and Shinji Watanabe, “End-to-end speech recognition with word-based rnn language models,” in IEEE SLT, 2018, pp. 389–396.
  • [76] V. Panayotov, G. Chen, D. Povey, and S. Khudanpur, “Librispeech: an ASR corpus based on public domain audio books,” in ICASSP, 2015, pp. 5206–5210.
  • [77] Zhong-Qiu Wang and DeLiang Wang, “A joint training framework for robust automatic speech recognition,” IEEE/ACM Transactions on Audio, Speech, and Language Processing, vol. 24, no. 4, pp. 796–806, 2016.
  • [78] Xuankai Chang, Wangyou Zhang, Yanmin Qian, Jonathan Le Roux, and Shinji Watanabe, “Mimo-speech: End-to-end multi-channel multi-speaker speech recognition,” in IEEE ASRU, 2019, pp. 237–244.
  • [79] Jian Wu, Zhuo Chen, Sanyuan Chen, Yu Wu, Takuya Yoshioka, Naoyuki Kanda, Shujie Liu, and Jinyu Li, “Investigation of practical aspects of single channel speech separation for asr,” in Interspeech, 2021, pp. 3066–3070.
  • [80] N. Condit-Schultz and D. Huron, “Catching the lyrics: intelligibility in twelve song genres,” Music Perception: An Interdisciplinary Journal, vol. 32, no. 5, pp. 470–483, 2015.
  • [81] Emir Demirel, Sven Ahlback, and Simon Dixon, “A recursive search method for lyrics alignment,” in MIREX, 2020.
  • [82] Xiaoxue Gao, Chitralekha Gupta, and Haizhou Li, “Lyrics transcription and lyrics-to-audio alignment with music-informed acoustic models,” in MIREX, 2021.
[Uncaptioned image] Xiaoxue Gao (Member, IEEE) received the Ph.D. degree from the National University of Singapore (NUS) in 2022 and received the B.Eng. degree in Electronic Information Science and Technology from Nanjing University, China, in 2017. She is currently a post-doctoral research fellow at NUS. Her research interests include automatic lyrics transcription, speech recognition, speech-to-singing conversion, singing information processing, music information retrieval and multi-modal processing.
[Uncaptioned image] Chitralekha Gupta (Member, IEEE) received the Ph.D. degree from the National University of Singapore (NUS) in 2019, Masters in engineering degree from Indian Institute of Technology Bombay (IIT-B), India in 2011, and Bachelors in engineering degree from Maharaja Sayajirao University (MSU) Baroda, India in 2008. She is currently a Senior Research Fellow at the School of Computing, NUS. She has been awarded the start-up grant of Graduate Research Innovation Program NUS and has founded the music tech company MuSigPro Pte. Ltd. in Singapore in 2019. She is the recipient of the NUS Dean’s Graduate Research Achievement Award 2018, the NUS School of Computing Innovation Prize 2018, and the Best Student Paper Award in APSIPA 2017. Her current research interests are neural audio synthesis, affective modeling of audio, singing voice analysis, and applications of ASR in music. She has played an active role in the organizing committees of several international conferences and challenges including ISMIR 2022, ICASSP 2022, and MIREX 2020.
[Uncaptioned image] Haizhou Li (Fellow, IEEE) received the B.Sc., M.Sc., and Ph.D. degrees in electrical and electronic engineering from the South China University of Technology, Guangzhou, China, in 1984, 1987, and 1990 respectively. He is currently a Presidential Chair Professor and the Executive Dean of the School of Data Science, The Chinese University of Hong Kong, Shenzhen, China. He is also an Adjunct Professor with the Department of Electrical and Computer Engineering, National University of Singapore, Singapore. Prior to that, he taught with the University of Hong Kong, Hong Kong, (1988–1990) and South China University of Technology, (1990–1994). He was a Visiting Professor with CRIN in France (1994–1995), Research Manager with the AppleISS Research Centre (1996–1998), Research Director with Lernout & Hauspie Asia Pacific (1999–2001), a Vice President with InfoTalk Corp. Ltd. (2001–2003), and the Principal Scientist and Department Head of Human Language Technology with the Institute for Infocomm Research, Singapore (2003–2016). His research interests include automatic speech recognition, speaker and language recognition, natural language processing. Dr. Li was an Editor-in-Chief of IEEE/ACM Transactions on Audio, Speech and Language Processing (2015–2018), a Member of the Editorial Board of Computer Speech and Language since 2012, an elected Member of IEEE Speech and Language Processing Technical Committee (2013–2015), the President of the International Speech Communication Association (2015–2017), the President of Asia Pacific Signal and Information Processing Association (2015–2016), and the President of Asian Federation of Natural Language Processing (2017–2018). He was the General Chair of ACL 2012, INTERSPEECH 2014, ASRU 2019 and ICASSP 2022. Dr. Li is a Fellow of the ISCA, and a Fellow of the Academy of Engineering Singapore. He was the recipient of the National Infocomm Award 2002, and the President’s Technology Award 2013 in Singapore. He was named one of the two Nokia Visiting Professors in 2009 by the Nokia Foundation, and U Bremen Excellence Chair Professor in 2019.