Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

work on other languages #16

Open
taalua opened this issue May 3, 2024 · 2 comments
Open

work on other languages #16

taalua opened this issue May 3, 2024 · 2 comments

Comments

@taalua
Copy link

taalua commented May 3, 2024

Hi,

For fine-tuning the current model to other languages, is it better to use the existing trained model and prompt tokenizer "parler-tts/parler_tts_mini_v0.1" or maybe it better train from scratch with a custom tokenizer? Any suggestions for the multilingual tokenizer if using espeak-ng? Thank you for your insights.

@ylacombe
Copy link
Collaborator

ylacombe commented May 9, 2024

Hey @taalua, it depends on the languages you want to fine-tune on!
If the flan T5 tokenizer covers your language (say Spanish or French), you can fine-tune the existing model, otherwise you probably need another custom tokenizer or one suited for multilinguality (say mt5 or something) and to train your model from scratch!

@thorstenMueller
Copy link

Hi @ylacombe , congrats for your impressive work 👏.

I created a german "Thorsten-Voice" dataset on Huggingface to be used for a Parler TTS training (https://huggingface.co/datasets/Thorsten-Voice/TV-44kHz-Full).

Right now i'm on my first step with "dataspeech" and ask myself if i have to or can simply adjust this code or have to switch to another phonemizer like "phonemizer" to support my work on a pure german single speaker voice dataset.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants