Replies: 1 comment
-
Hi Ahmet, thank you very much for the kind words! The trouble with higher dimensions is two-fold:
This should answer (1), now the other queries:
Cheers! |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
First of all, this work has been inspiring for me to say the least and I appreciate the structuring of the repos and the work on tcnn!!!
I have been tinkering with tcnn and multi-res hash grid encodings with higher input dimensions.
I had to enable 5/6/7 dimensions here:
https://github.com/NVlabs/tiny-cuda-nn/blob/bd29d77c589e7593680a3aa508e35a0136df904b/include/tiny-cuda-nn/encodings/grid.h#L931
I am wondering why they are disabled by default. My experience so far has been that it may be causing more hash map collusions but I didn't dig deeper to confirm.
I am also curious to confirm my suspicion that most of the representation may be encoded in the hash grid resolutions (especially for my case, where I use this as a material encoder in UV space) rather than the following MLP
where the input and the data structure lookup is concatenated.(apparently encoding is not concatenated as I just found out here #151 (comment), but my question still may be relevant)Has there been experimental results where having more learnable features in the hash grid improved convergence times?
I have had faster convergence with 3e-4 learning rate and bigger log2 hash map sizes so far but I usually can't fit my structures into L1/L2 caches where the Fully Fused MLP shines... I am curious, if there may be a sweet spot where the convergence is fast enough at lower amounts of epochs with some hyper parameter tuning while sacrificing the structural packing and fusing optimizations.
summary of my questions (which are all over the place):
1.) Why higher dimensional inputs for hash grid encodings are disabled by default?
2.) Are there any best practices or observations for hash collusions?
3.) Are these representations memorized mostly in the spatial data structures rather than the MLP and perhaps, has there been any results to confirm or deny that?
4.) At the expense of not having some optimizations, can there be faster convergence, and if there has been any experiments for that?
Thanks for your time in advance and the great work \o/
Beta Was this translation helpful? Give feedback.
All reactions