Skip to content

Are there non-obvious best practices for saving memory and operations #1482

Discussion options

You must be logged in to vote

I think IncompressibleModel{MultiGPU} might still be a little far off since we need distributed FFT support for GPUs from PencilFFTs.jl. Could happen soon but no ETA. ShallowWaterModel{MultiGPU} does work thanks to @francispoulin but might need some profiling to find bottlenecks.

I think you touched on the obvious ways!

I guess another way is to find a bigger GPU. Some of the higher-end NVIDIA GPUs have 32 GB of memory, but not sure there are any common ones with more.

A riskier way is to use Float32 to half your memory footprint, but then you might end up having to manage truncation errors as discussed in #1410.

I assume you're already using advection = WENO5() but you could use a higher…

Replies: 2 comments 9 replies

Comment options

You must be logged in to vote
6 replies
@tomchor
Comment options

tomchor Mar 16, 2021
Collaborator Author

@glwagner
Comment options

@tomchor
Comment options

tomchor Mar 16, 2021
Collaborator Author

@ali-ramadhan
Comment options

@glwagner
Comment options

Answer selected by tomchor
Comment options

You must be logged in to vote
3 replies
@tomchor
Comment options

tomchor Mar 16, 2021
Collaborator Author

@glwagner
Comment options

@tomchor
Comment options

tomchor Mar 16, 2021
Collaborator Author

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
3 participants