-
Notifications
You must be signed in to change notification settings - Fork 93
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
How to build the validation data? #62
Comments
What is the "G identity mapping loss" step? |
Thank you so much for a quick response! Let me play around with it. The training part is working, but evaluation part is failing. The G identity mapping loss step is this part code that does something with the train data in Solver.train method:
|
I've fixed my issue. The thing was I was creating the speaker embeddings as a 1 dimensional array, while the model expected 2 dimensional. So I have 1085 speakers and for each speaker I created one-hot encoding vector of the size [1085]. While the model expected a vector of the size [1, 1085]. Thank you again for your help! |
|
Hello,
thank you so much for the code and paper! I'm trying to train the model on speech command data. I've made the train and validation data sets through 2 scripts:
make_spect_f0.py
andmake_metadat.py
, but the model fails on the validation step, on this line :x_identic_val = self.G(x_f0, x_real_pad, emb_org_val)
The error is:
RuntimeError: The expanded size of the tensor (192) must match the existing size (1085) at non-singleton dimension 1. Target sizes: [-1, 192, -1]. Tensor sizes: [1085, 1]
.I'm not sure why there is a mismatch as
self.G
worked. Although there is the "G identity mapping loss" step which preprocess the input before feeding toself.G
. Do I need to do the same with the validation data? Also 192 is themax_len_pad = 192
, while 1085 is the number of the speakers (dim_spk_emb = 1085
). Do I need to change the max_len_pad?I'll appreciate for any help or direction!
My hparams.py is below
The text was updated successfully, but these errors were encountered: