Replies: 4 comments 5 replies
-
The loss here looks actually rather strange; can you give us some info about the data size and batch size? And the model type and dimensions need to be carefully considered based on the data you have, so depending on how you sweep this, you could very much overfit ... Imagine data size of 10, and dim >9 would not work; or a model with receptive field of >10 with length of data <10... would not work, etc. 24 hidden units also seems quite small, but again depends on input size Can you just us some more details, and also a visualization of the embedding would be helpful as well. I would set temp=1 to start. |
Beta Was this translation helpful? Give feedback.
-
Beta Was this translation helpful? Give feedback.
-
no problem :) happy to help! okay so "either lines or unstructured clouds/spheres" means that it's collapsing for sure; if you don't use a 1D label to start, but rather unsupervised (CEBRA-Time), can you get embeddings with more structure? :D |
Beta Was this translation helpful? Give feedback.
-
Beta Was this translation helpful? Give feedback.
-
Hi,
Trying CEBRA for the first time, basically trying to optimize decoding of a continuous variable. Tried a random search of the following hyperparameters and actually consistently getting negative decoding R^2 (e.g. -0.3, i.e. MSE is significantly higher than the label variance).
Also tried to include
temperature
as a hyper-param (withtemperature_mode=constant
) but that somehow shortens the embedding from 37,200 samples to 9,300 so that the decoder training led to the following error:ValueError: Invalid shape: y and X must have the same number of samples, got y:37200 and X:9300.
Note: baseline models (e.g. XGBoost on single time-points with no temporal context) perform reasonably well on this task (R^2 typically
between 0.05-0.25)
Any ideas???
Thanks!
Beta Was this translation helpful? Give feedback.
All reactions