Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

about n_spks #30

Open
0913ktg opened this issue Feb 19, 2024 · 6 comments
Open

about n_spks #30

0913ktg opened this issue Feb 19, 2024 · 6 comments

Comments

@0913ktg
Copy link

0913ktg commented Feb 19, 2024

Hello p0p4k,

I am deeply grateful for the code you have provided.

I have a question while adapting it to a Korean version. I am preparing to use a speech-to-text dataset with approximately 2000 speakers.

However, the dataset does not contain speaker labels for the voice data. From what I understand in the paper, it seems that speaker information is learned using a 3-second prompt without explicitly using speaker labels.

If my understanding is correct, this suggests that speaker labels are not necessary for training the model, and thus, not required in the filelist for multi-speaker synthesis.

Yet, I noticed that the code is designed to use speaker labels at the first index of the filelist when n_spks is greater than 1.

I would be extremely grateful if you could clarify this part for me.

My understanding of this paper is still quite limited, and I apologize if my question seems naive.

Thank you.

@p0p4k
Copy link
Owner

p0p4k commented Feb 19, 2024

Just ignore the speaker list or anything regarding n_spks. Send audio text to the dataloader as if it's one big dataset. Pflow will take care of it. Your understanding is correct. Nspks exists for future use.

@0913ktg
Copy link
Author

0913ktg commented Feb 20, 2024

Thank you for your response.
Have a great day.

@0913ktg
Copy link
Author

0913ktg commented Mar 1, 2024

Hello p0p4k,

I've been testing single-speaker training using your pflow code. Unfortunately, due to poor data quality and training multi-speaker data with n_spks set to 1, the mel-spectrogram was not generated effectively.

I plan to experiment with a multi-speaker dataset that includes spk_id labels, hoping to train a robust Korean model. The preprocessing has begun, and the training is underway.

Additionally, I noticed you forked the naturalspeech2 code. I'm curious if you have any development plans related to it. I'm quite interested in training with naturalspeech2 myself.

Thank you.

@p0p4k
Copy link
Owner

p0p4k commented Mar 1, 2024

  • I want to add text|audio pair in the ns2 repo for training it and see if it works.
  • In my head, I think for model like pflow even if data is poor, we can augment the pitch and generate infinite data for same text (spoken content is same, but pitch change will change the style). Since it doesnt take in spk_id and takes in a prompt, we should be able to train it with even a single speaker to generate any voice based on the prompt style. What do you think about this?

@0913ktg
Copy link
Author

0913ktg commented Mar 4, 2024

I changed the dataset and experimented again.
It didn't generate mel-spectrograms well on the old data, but it seems to generate them almost as well on this data.
I think the previous data was recorded conversations of normal people, so the text and speech were inconsistent and the speaker's speaking style was too strong.
The new data is much less noisy, the pronunciation is clearer, and the speakers are consistent and faithful in their recordings.

image

@p0p4k
Copy link
Owner

p0p4k commented Mar 4, 2024

Good news! Thanks for update, train for longer, 그리고 샘플을 공유 해주세요.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants