CEBRA Interpretability #143
Closed
xanderladd
started this conversation in
General
Replies: 1 comment 3 replies
-
Hi @xanderladd , great question! Please have a look at this earlier discussion on this topic, we are actively working on it. Here is our latest work on the topic which we will integrate into CEBRA: https://sslneurips23.github.io/paper_pdfs/paper_80.pdf |
Beta Was this translation helpful? Give feedback.
3 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Thanks for this tool and for the support you give on github!
I am able to produce embeddings of neural data that have a clear interpretation in terms of behavior. But I am wondering if there is a way to do something like PCA loadings in order to understand the contribution of an individual neuron to the embeddings. I know this is a tall order because the weights of the neural network are not intuitively interpretable. Is this worth thinking about? Or should I just consider the positive / negative sample neural nets to be black boxes?
Some ideas I had for this were:
Ultimately, what would be great to know is some way of quantifying a single neuron (or a subset of neurons) contribution to the embeddings. I don't expect to make strong causal arguments this way, but correlational evidence for further hypothesis discovery would be good.
Beta Was this translation helpful? Give feedback.
All reactions