PyGeM taking up a lot of memory #274
BilalAnwar165
started this conversation in
General
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
Hi,
I have read in documentation of PyGeM that it has been tested up to mesh size of 14 million elements. Curently, I am using PyGeM for a mere 2-3 million of elements, but it is taking up a lot of memory up to the extent that it becomes practically impossible to provide that much RAM. Because of this, it get stuck a when it calls rbf.call method to deform the volume mesh. I want to know is it expected so, or am I doing something wrong?
For your info, for the above mesh size , it asks for more than 27.3GiB of Memory otherwise it either kills the process or shows the error
"_H = np.zeros((npts + 3 + 1, npts + 3 + 1))
numpy.core._exceptions.ArrayMemoryError: Unable to allocate 27.3 GiB for an array with shape (60558, 60558) and data type float64"
The problem is I am just testing this tool on this small mesh. the original mesh has a size of 8-10 Million elements. In that case, It is going to require a huge amount of RAM, which is practically not possible.
Do you have any solution or input regarding this problem?
Beta Was this translation helpful? Give feedback.
All reactions