You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
GPU memory is not properly freed when switching to other meshes, eventually leading to CUSPARSE_STATUS_ALLOC_FAILED:
Traceback (most recent call last):
File "scripts/show_largesteps_memory_leak.py", line 16, in <module>
v = from_differential(M, u, 'Cholesky')
File "/home/xuzhen/miniconda3/envs/flame/lib/python3.8/site-packages/largesteps/parameterize.py", line 51, in from_differential
solver = CholeskySolver(L)
File "/home/xuzhen/miniconda3/envs/flame/lib/python3.8/site-packages/largesteps/solvers.py", line 130, in __init__
self.solver_1 = prepare(self.L, False, False, True)
File "/home/xuzhen/miniconda3/envs/flame/lib/python3.8/site-packages/largesteps/solvers.py", line 68, in prepare
_cusparse.scsrsm2_analysis(
File "cupy_backends/cuda/libs/cusparse.pyx", line 2103, in cupy_backends.cuda.libs.cusparse.scsrsm2_analysis
File "cupy_backends/cuda/libs/cusparse.pyx", line 2115, in cupy_backends.cuda.libs.cusparse.scsrsm2_analysis
File "cupy_backends/cuda/libs/cusparse.pyx", line 1511, in cupy_backends.cuda.libs.cusparse.check_status
cupy_backends.cuda.libs.cusparse.CuSparseError: CUSPARSE_STATUS_ALLOC_FAILED
To reproduce, run this code example with this example mesh (extract armadillo.npz and place it where you run the code below):
Thank you for this report. We replaced cusparse by cholespy for the solver part, so this should not be a problem anymore. nrhs is also not fixed to 3 anymore (up to 128 on the GPU)
GPU memory is not properly freed when switching to other meshes, eventually leading to
CUSPARSE_STATUS_ALLOC_FAILED
:To reproduce, run this code example with this example mesh (extract
armadillo.npz
and place it where you run the code below):While running the code above, you should see the GPU memory continuously increase but the expected behavior is that it stays constant.
For example, the result of
nvidia-smi dmon -s m
while running the code should be something like:The text was updated successfully, but these errors were encountered: