PyGeM taking up a lot of memory #274
BilalAnwar165
started this conversation in
General
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Hi,
I have read in documentation of PyGeM that it has been tested up to mesh size of 14 million elements. Curently, I am using PyGeM for a mere 2-3 million of elements, but it is taking up a lot of memory up to the extent that it becomes practically impossible to provide that much RAM. Because of this, it get stuck a when it calls rbf.call method to deform the volume mesh. I want to know is it expected so, or am I doing something wrong?
For your info, for the above mesh size , it asks for more than 27.3GiB of Memory otherwise it either kills the process or shows the error
"_H = np.zeros((npts + 3 + 1, npts + 3 + 1))
numpy.core._exceptions.ArrayMemoryError: Unable to allocate 27.3 GiB for an array with shape (60558, 60558) and data type float64"
The problem is I am just testing this tool on this small mesh. the original mesh has a size of 8-10 Million elements. In that case, It is going to require a huge amount of RAM, which is practically not possible.
Do you have any solution or input regarding this problem?
Beta Was this translation helpful? Give feedback.
All reactions