Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Impact of RoPE and Adding Prompts in Cross Attention on Model Performance #106

Open
LiuZH-19 opened this issue Aug 15, 2024 · 0 comments
Open

Comments

@LiuZH-19
Copy link

Hello,

I noticed that in the recent architecture improvements, modules for RoPE positional encoding and adding Prompts in Cross Attention were included. However, it seems that the newly released two Parler-TTS checkpoints did not utilize these features (if I understood correctly). Do you have any ablation study results on the impact of using RoPE positional encoding and adding Prompts in Cross Attention? I’m interested in understanding how each of these modules affects the final model performance.

Additionally,is there a plan to update the training guide for the latest checkpoints? I’m particularly keen on learning how to fine-tune the new checkpoints.

Thank you for your amazing work!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant