YourTTS and voice cloning #291
Replies: 3 comments 5 replies
-
Yes, for multi-speaker models you always need to specify a speaker to do TTS, Coqui doesn't have a concept of a default speaker for a model. The Fairseq models are single-speaker, so it's not needed there. But note that YourTTS specifically also supports voice cloning directly and you will probably achieve better results like this instead of separately doing voice conversion afterwards. And in that case there's no need to specify a from TTS.api import TTS
tts = TTS('tts_models/multilingual/multi-dataset/your_tts')
tts.tts_to_file('hello world', speaker_wav='reference.wav', language='en') |
Beta Was this translation helpful? Give feedback.
-
ok got it thanks! |
Beta Was this translation helpful? Give feedback.
-
Could you confirm yourtts is only eng, fr-fr, pt-br? I saw somewhere it was supporting more languages, but how? |
Beta Was this translation helpful? Give feedback.
-
I'm using tts_with_vc() with yourtts like this:
tts.tts_with_vc(
text='hello world'],
speaker='male-en-2',
speaker_wav=voice_clone_path,
language='en'
)
I had to add speaker= to make it work, although with fairseq I don't need to add it. is it correct? btw the result is not really the cloned voice but a voice sounding like the builtin voice with a very little sound from the voice_clone. is it normal? or does it need more settings? thanks
Beta Was this translation helpful? Give feedback.
All reactions