Mellotron: Multispeaker expressive voice synthesis by conditioning on rhythm, pitch and global style tokens

10/26/2019 ∙ by Rafael Valle, et al. ∙ 0

Mellotron is a multispeaker voice synthesis model based on Tacotron 2 GST that can make a voice emote and sing without emotive or singing training data. By explicitly conditioning on rhythm and continuous pitch contours from an audio signal or music score, Mellotron is able to generate speech in a variety of styles ranging from read speech to expressive speech, from slow drawls to rap and from monotonous voice to singing voice. Unlike other methods, we train Mellotron using only read speech data without alignments between text and audio. We evaluate our models using the LJSpeech and LibriTTS datasets. We provide F0 Frame Errors and synthesized samples that include style transfer from other speakers, singers and styles not seen during training, procedural manipulation of rhythm and pitch and choir synthesis.



There are no comments yet.


page 4

This week in AI

Get the week's most popular data science and artificial intelligence research sent straight to your inbox every Saturday.

1 Introduction

Speech synthesis is typically formulated as the conversion of text to speech (TTS). This formulation, however, leaves out control for all the aspects of speech not contained in the text. Here we approach the problem of expressive speech synthesis which includes not just text, but other characteristics such as pitch, rhythm and emphasis. There are formulations to expressive speech synthesis that require animated and emotive voice data. This is an inconvenient drawback given the limited access to such data. In our approach, we can make a voice emote and sing without any such data.

Recent approaches that utilize deep learning for expressive speech synthesis combine text and a learned latent embedding for prosody or global style [19, 20]. While these approaches have shown promise, manipulating such latent variables only offers a coarse control over expressive characteristics of speech. Mellotron was motivated by the desire for fine grained control over these expressive characteristics. Notably, we show that it is easy to condition Mellotron on pitch and rhythm information automatically extracted from an audio signal or music score.

By accounting for melodic information such as pitch and rhythm, expressive speech synthesis with Mellotron can be easily extended to singing voice synthesis (SVS) [13, 8]. Unfortunately, recent attempts [8] require a singing voice dataset and heavily quantized pitch and rhythm data obtained from a digital representation of a music score, for example MIDI [11] or musicXML [5]. Mellotron does not require any singing voice in the dataset nor manually aligned pitch and text in order to synthesize singing voice.

Mellotron can make a voice emote and sing without emotion or singing data. Training Mellotron is very simple and only requires read speech and transcriptions. During inference, we can change the generated voice’s speaking style, make it emote or sing by extracting pitch and rhythm characteristics from an audio file or a music score. As a bonus, with Mellotron we can explore latent characteristics from an audio corpus by sampling a dictionary of learned latent characteristics. In summary, Mellotron is a versatile voice synthesis111Includes speech synthesis, singing voice synthesis, etc. model that enables the combination of characteristics from different sources and generalizes to characteristics not seen in training data.

2 Method

Mellotron is a voice synthesis model that uses a combination of explicit and latent variables. Whereas well-established signal processing algorithms provide explicit variables that are valuable to expressive speech such as fundamental frequency contours and voicing decisions, deep learning strategies can be used to learn latent variables that express characteristics of an audio corpus that are unknown to the user and hard to formalize.

We factorize a single speaker mel-spectrogram

into explicit variables such as text, speaker identity, a fundamental frequency contour augmented with voiced/unvoiced decisions and two latent variables learned by the model during training. The first latent variable refers to a dictionary of vectors that can be queried with an audio input or sampled directly as described in

[20]. The second latent variable is the learned attention map between the text and the mel-spectrogram as described in [18].

From now on we will refer to the augmented fundamental frequency contour as pitch contour and refer to the first and second latent variables as global style tokens (GST) and rhythm respectively.

We are interested in factorizing , where T represents the text, S represents the speaker identity, P represents the pitch contour, R represents the rhythm and Z represents the global style tokens. Given this formulation, during training we maximize the following:


where the superscript represents the i-th mel, , and represent the text, speaker, and pitch contour associated with the i-th mel, represents the learned alignments between the text and mel-spectrogram frames, represents the global style token conditioned on as presented in [20], and represents the model parameters.

The explicit factors offers two advantages. First, by providing the model with text and speaker information, we prevent the problem of entanglement between text and speaker information. Second by providing the model with pitch contour and voicing information, we are able to directly control pitch and voicing decisions during inference.

Similarly the latent factors offers two advantages. First, by learning the alignment map between the text and mel-spectrogram during training, we do not need to extract phoneme alignments for training and can control the rhythm during inference by providing the model with an alignment map. Second, by providing the model with a dictionary of latent variables, we are able to learn latent factors that are harder to express or extract explicitly, thus leveraging the full power of latent variables.

Using this formulation we are able to transfer the text, rhythm and pitch contour from a source, e. g. audio signal or musical score, to a target speaker by replacing the variables in Equation 1 accordingly. For example, we first collect the text, pitch and rhythm (), from the source, sample a GST from the GST dictionary learned by Mellotron, and chose a target speaker .


should now have the same text, pitch and rhythm as the source, latent characteristics obtained from the global style token and the voice of the target speaker. In our current formulation, the target speaker, , would always be found in the training set, while the source text, pitch and rhythm () could be from outside the training set. This allows us to train a model that makes a voice emote and sing without using any singing voice in the training dataset, without any manual labelling of emotions nor pitch, and without any manual alignments between words and audio, nor between pitch and audio.

3 Implementation

In this section we are going to describe our model architecture and our training and inference setups. We plan to release our implementation and pre-trained models on github.

3.1 Architecture

Mellotron extends Tacotron 2 GST [20] with speaker embeddings and pitch countours. Unlike [4, 15]

, where site specific speaker embeddings are used, we use a single speaker embedding that is channel-wise concatenated with the encoder outputs over every token. The pitch contour goes through a single convolution layer followed by a ReLU non-linearity. We experiment with kernel sizes 1 and 3 and convolution dimensions 1 and 8. The pitch contour is channel-wise concatenated with the decoder inputs. We use phoneme representations whenever possible.

3.2 Training

Our implementation only requires text and audio pairs with a speaker id. Our pitch contours are automatically extracted using the Yin algorithm [3] with harmonicity thresholds between 0.1 and 0.25. Unlike [8], during training our model does not require manually aligned text, pitch and mel-spectrogram. We use the L2 loss between ground truth and predicted mels described in [20] without any modifications.

3.3 Inference

Following the description in Section 2, during inference we provide Melloron with text, rhythm and pitch information that is obtained either from an audio signal or from a musical score, a global style token and a speaker id.

3.3.1 Audio Signal

Obtaining text, rhythm and pitch information consists of three steps. First, we extract text information from an audio file by either using an automatic speech recognition model [9, 1] or by manually transcribing the text. The text information is pre-processed with our text cleaners and then converted from graphemes to phonemes.

Second, we extract rhythm information by using a forced-alignment tool [14, 10] or by using Mellotron as a forced-aligner. Alignment maps can be obtained with Mellotron by performing a teacher-forced forward pass using the data from the source signal. Whenever necessary, we fine tune the alignment maps by hand or by training Mellotron on the source signal for a few iterations with small learning rate.

The pitch data is obtained by using Yin [3] or Melodia [17]. In our quantitative experiments we use Yin to replicate the setup described in [19]. In our qualitative experiments we use Melodia instead as we find it to be more precise than Yin, specially with regards to false voiced decisions.

3.3.2 Music Score

We operate on music scores in XML format containing event tuples with pitch, note duration and syllables for each part in the score. We directly convert pitch to frequency and use the FFT hop size to convert event durations from seconds to frames. We remind the reader that although we refer to pitch, our model’s representation of pitch is continuous.

We concatenate the syllables into words and convert graphemes to phonemes. For single phone events, the duration of each phone is equal to the duration of the event. For multi-phone events, the duration of each phone is dependent on its type: we use heuristics to assign durations between 20 and 100ms to consonants and assign the remainder of the event’s duration to vowels. For example, consider a one second long single note event on the word

Bass with phoneme representation [B, AE, S]. We set B to 20 ms, S to 100 ms and the remaining duration to AE, and hence have full control over the duration of each phone.

4 Experiments

We train our models using the LJSpeech (LJS) dataset [6], the Sally dataset, a proprietary single speaker dataset with 20 hours, and a subset of LibriTTS [21]. All datasets used in our experiments are from read speech.

We provide results that include style transfer222Transferring text, rhythm and pitch contour to a target speaker. from source speakers seen and unseen in the dataset, from singers, procedural manipulation of rhythm and choir synthesis from music scores. Visit our website333 to listen to Mellotron samples.

4.1 Training Setup

For all the experiments, we trained on LJS, Sally and the train-clean-100 subset of LibriTTS with over 100 speakers and 25 minutes on average per speaker. Speakers with less than 5 minutes of data and files that are larger than 10 seconds were filtered out. We do not perform any data augmentation, hence any extension to a speaker’s characteristics such as vocal range and speech rate is made possible with Mellotron.

We use a sampling rate of 22050 Hz and mel-spectrograms with 80 bins using librosa mel filter defaults. We apply the STFT with a FFT size of 1024, hop size of 256, and window size of 1024 samples.

We use the ADAM [7] optimizer with default parameters, start with a 1e-3 learning rate and anneal the learning rate as the loss starts to plateau. We decrease training time by using a single NVIDIA DGX-1 with 8 GPUs.

For decoding the mel-spectrograms produced by Mellotron, we use a single WaveGlow [16] model trained on the Sally dataset. Our results suggest that Waveglow can be used as an universal decoder.

In our setup, we find it easier to first learn attention alignments on speakers with large amounts of data and then fine tune to speakers with less data. Thus, we first train Mellotron on LJS and Sally and finetune it with a new speaker embedding on LibriTTS, starting with a learning rate of 5e-4 and annealing the learning rate as the loss starts to plateau.

4.2 Quantitative Results

In this section we provide quantitative results that compare Gross Pitch Error (GPE) [12], Voicing Decision Error (VDE) [12] and F0 Frame Error (FFE) [2] between Mellotron and E2E-Prosody [19]. Following [19], all pitch and voicing metrics are computed using the Yin algorithm [3]

. Due to the rhythm conditioning, our reference and predicted audio have the same length and does not require padding.

The results in Table 1 below show that by conditioning on pitch we can drastically reduce the error between the source and the synthesized voice. For singing voice, low pitch error is extremely important otherwise the melody might lose its identity. For prosody transfer, a lower FFE provides evidence that the style will be more precisely transferred to the target.

Model Voice GPE VDE FFE
E2E-Prosody Single - -
E2E-Prosody Multi - -
Mellotron LJS-Sally 0.08% 9.19%
Mellotron LibriTTS 0.08% 8.69%
Table 1: GPE, VDE, FFE for Mellotron and E2E-Prosody. The reference is always the same speaker.

4.3 Style transfer from Audio Signal

Mellotron is able to emote and match the style of an input audio by replicating its rhythm or both its rhythm and pitch. Overall, we note that our experiments using audio data are directly impacted by the quality of the rhythm and pitch contours provided to the model. Whereas Melodia provides rather precise pitch contours, we find that the rhythm data obtained from forced-alignments had to be constantly fine-tuned. In all audio experiments we obtain the rhythm by fine-tuning alignment maps obtained by using Mellotron as a forced-aligner. Occasionally we find that some of the pitch contours seem to be outside of a speaker’s vocal range. When this happens, Mellotron defaults to a constant highest or lowest pitch value. We circumvent this by scaling the pitch contour by a constant to matches the speaker’s vocal range.

4.3.1 Rhythm Transfer

In this experiment we transfer the rhythm and its associated text from a source audio signal to a target speaker. Our formulation provides procedural control over the duration of every phoneme, hence allowing for simple manipulations such as changing the speech rate or complex effects like speeding up or slowing down. In rhythm transfer, we provide Mellotron with an array of zeros as the pitch contour.

We show examples where we transfer the rhythm from an excerpt by Nicki Minaj to Sally. We showcase the procedural capabilities of Mellotron by processing the source rhythm with a function that produces an accelerando starting at half the speed and accelerating to twice the speed. For comparison, we also provide samples conditioned on the pitch contour from Nicki’s track. Figure 1 shows the alignment maps.

Figure 1: Left: source alignment. Right: processed alignment.

4.3.2 Rhythm and Pitch Transfer

By conditioning on both rhythm and pitch, we can express characteristics of the source speaker’s style. An interesting application is the creation of a hybrid with the style from a source speaker but the voice from another speaker. We show an example where we transfer the characteristics of a solemn speech to Sally. We see that Mellotron contains the same pauses and speech rate as the source which adds to the solemnity of the speech. For comparison, we provide the same phrases synthesised with the original Tacotron 2 which fails to convey the same solemnity.

4.4 Singing Voice Synthesis

Mellotron is able to generalize to rhythm and pitch from styles and speakers not in the training set. We are able to synthesize singing voice from a wide range of input speakers across a range of music styles such as rap, pop, Hindustani and western European classical music.

4.4.1 Singing Voice from Audio Signal

Figure 2 shows an example where we use the Sweet Dreams sample from the E2E-Prosody paper [19] and transfer its text, rhythm and scaled pitch to Sally. Figure 2 shows that Mellotron’s pitch contour is closer to the source than E2E-Prosody is.

Figure 2: Source, Mellotron and E2E-Prosody pitch contours.

4.4.2 Style transfer from Music Score

Unlike the experiments on audio, the rhythm and pitch contours provided to the model by a music score are correct by design. We provide a 4-part example with 20 voices per part on an excerpt of Handel’s Hallelujah, a 8-part example with 1 voice per part on Ligeti’s Lux Aeterna and a single voice example synthesizing the opening flute intro from Debussy’s Prélude à l’après-midi d’un faune. Except from cases where the pitch is beyond the speaker’s vocal range, such as in Handel’s sample, Mellotron has very precise pitch and rhythm.

5 Conclusion

In this paper we described Mellotron, a multispeaker voice synthesis model that allows for direct control of style by conditioning on rhythm and pitch obtained from an audio signal or a music score.

Our numerical results show that Mellotron is superior to other models with respect to F0 Frame Error. Our qualitative results show that Mellotron is able to generate speech in a variety of styles ranging from read speech to expressive speech, from slow drawls to rap, and from monotonous voice to singing voice although none of these styles are present in the training data.

Recent singing voice synthesis papers [8] state that ”even in the case of a real recording sample recorded by listening to the original midi accompaniment, it is not easy to adjust the timing and pitch of the correct note” indicating that it is difficult for professional human singers and synthesized voice to match a source audio or source music score perfectly. Our results show that one of the advantages of Mellotron is that the rhythm and pitch contour of a synthesized sample is extremely similar to the source audio file or music score, under the assumption that the pitch is within a speaker’s vocal range. When outside a speaker’s vocal range, Mellotron defaults to either the lowest tone or highest tone.

For future work, we plan to study the effect of rhythm and pitch contours on the audio quality by comparing samples conditioned on pitch and rhythm data obtained from audio signals versus music scores. With respect to pitch, we are also interested in understanding the effect of multi-speaker training on a speaker’s vocal range and extending a speaker’s vocal range as much as possible. Last, we would like to train Mellotron on a animated and emotive storytelling style dataset to investigate the contribution of such dataset to Mellotron.


  • [1] C. Chiu, T. N. Sainath, Y. Wu, R. Prabhavalkar, P. Nguyen, Z. Chen, A. Kannan, R. J. Weiss, K. Rao, E. Gonina, et al. (2018) State-of-the-art speech recognition with sequence-to-sequence models. In 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 4774–4778. Cited by: §3.3.1.
  • [2] W. Chu and A. Alwan (2009) Reducing f0 frame error of f0 tracking algorithms under noisy conditions with an unvoiced/voiced classification frontend. In 2009 IEEE International Conference on Acoustics, Speech and Signal Processing, pp. 3969–3972. Cited by: §4.2.
  • [3] A. De Cheveigné and H. Kawahara (2002)

    YIN, a fundamental frequency estimator for speech and music

    The Journal of the Acoustical Society of America 111 (4), pp. 1917–1930. Cited by: §3.2, §3.3.1, §4.2.
  • [4] A. Gibiansky, S. Arik, G. Diamos, J. Miller, K. Peng, W. Ping, J. Raiman, and Y. Zhou (2017) Deep voice 2: multi-speaker neural text-to-speech. In Advances in neural information processing systems, pp. 2962–2970. Cited by: §3.1.
  • [5] M. Good (2001) MusicXML for notation and analysis. The virtual score: representation, retrieval, restoration 12, pp. 113–124. Cited by: §1.
  • [6] K. Ito (2017) The lj speech dataset. Note: Cited by: §4.
  • [7] D. P. Kingma and J. Ba (2014) Adam: a method for stochastic optimization. arXiv preprint arXiv:1412.6980. Cited by: §4.1.
  • [8] J. Lee, H. Choi, C. Jeon, J. Koo, and K. Lee (2019) Adversarially trained end-to-end korean singing voice synthesis system. arXiv preprint arXiv:1908.01919. Cited by: §1, §3.2, §5.
  • [9] J. Li, V. Lavrukhin, B. Ginsburg, R. Leary, O. Kuchaiev, J. M. Cohen, H. Nguyen, and R. T. Gadde (2019) Jasper: an end-to-end convolutional neural acoustic model. arXiv preprint arXiv:1904.03288. Cited by: §3.3.1.
  • [10] M. McAuliffe, M. Socolof, S. Mihuc, M. Wagner, and M. Sonderegger (2017) Montreal forced aligner: trainable text-speech alignment using kaldi.. In Interspeech, pp. 498–502. Cited by: §3.3.1.
  • [11] R. A. Moog (1986) MIDI: musical instrument digital interface. Journal of the Audio Engineering Society 34 (5), pp. 394–404. Cited by: §1.
  • [12] T. Nakatani, S. Amano, T. Irino, K. Ishizuka, and T. Kondo (2008-03) A method for fundamental frequency estimation and voicing decision: application to infant utterances recorded in real acoustical environments. Speech Communication 50 (3), pp. 203–214. External Links: ISSN 0167-6393, Link, Document Cited by: §4.2.
  • [13] M. Nishimura, K. Hashimoto, K. Oura, Y. Nankaku, and K. Tokuda (2016)

    Singing voice synthesis based on deep neural networks

    In Interspeech 2016, pp. 2478–2482. External Links: Document, Link Cited by: §1.
  • [14] Gentle forced aligner External Links: Link Cited by: §3.3.1.
  • [15] W. Ping, K. Peng, A. Gibiansky, S. O. Arik, A. Kannan, S. Narang, J. Raiman, and J. Miller (2017) Deep voice 3: scaling text-to-speech with convolutional sequence learning. arXiv preprint arXiv:1710.07654. Cited by: §3.1.
  • [16] R. Prenger, R. Valle, and B. Catanzaro (2019) Waveglow: a flow-based generative network for speech synthesis. In ICASSP 2019-2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 3617–3621. Cited by: §4.1.
  • [17] J. Salamon and E. Gómez (2012) Melody extraction from polyphonic music signals using pitch contour characteristics. IEEE Transactions on Audio, Speech, and Language Processing 20 (6), pp. 1759–1770. Cited by: §3.3.1.
  • [18] J. Shen, R. Pang, R. J. Weiss, M. Schuster, N. Jaitly, Z. Yang, Z. Chen, Y. Zhang, Y. Wang, R. Skerrv-Ryan, et al. (2018) Natural tts synthesis by conditioning wavenet on mel spectrogram predictions. In 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 4779–4783. Cited by: §2.
  • [19] R. Skerry-Ryan, E. Battenberg, Y. Xiao, Y. Wang, D. Stanton, J. Shor, R. J. Weiss, R. Clark, and R. A. Saurous (2018) Towards end-to-end prosody transfer for expressive speech synthesis with tacotron. arXiv preprint arXiv:1803.09047. Cited by: §1, §3.3.1, §4.2, §4.4.1.
  • [20] Y. Wang, D. Stanton, Y. Zhang, R. Skerry-Ryan, E. Battenberg, J. Shor, Y. Xiao, F. Ren, Y. Jia, and R. A. Saurous (2018) Style tokens: unsupervised style modeling, control and transfer in end-to-end speech synthesis. arXiv preprint arXiv:1803.09017. Cited by: §1, §2, §2, §3.1, §3.2.
  • [21] H. Zen, V. Dang, R. Clark, Y. Zhang, R. J. Weiss, Y. Jia, Z. Chen, and Y. Wu (2019) LibriTTS: a corpus derived from librispeech for text-to-speech. arXiv preprint arXiv:1904.02882. Cited by: §4.