Get the warning: W external/xla/xla/service/hlo_rematerialization.cc:2194] Can't reduce memory use below 10.96GiB (11774361600 bytes) by remateri alization; only reduced to 11.62GiB (12478088656 bytes)
#15681
Replies: 3 comments
-
I met exactly the same problem on my RTX3090. |
Beta Was this translation helpful? Give feedback.
-
I have the same issue, using jax However I also get the following error message:
I wonder whether having an old cuda version may be causing the re-materialization issue |
Beta Was this translation helpful? Give feedback.
-
I have the same problem. My GPU is always using 18607MiB / 24564MiB even if I reduce the size of my simulation (eg number of samples). I am using the following versions of things:
I also have the following environmental variables set:
I never get this problem on my other PC, which runs Ubuntu 22.04. |
Beta Was this translation helpful? Give feedback.
-
I got the following warning when benchmarking Hugging Face transformers' Bert model.
It seems that XLA is trying to reduce memory usage to 10.96GB by rematerialization, but my GPU has 16GB VRAM. I also set
XLA_PYTHON_CLIENT_MEM_FRACTION=.97
andnvidia-smi
shows that JAX pre-allocate ~15GB VRAM. So why does it need to rematerialize? Would it hurt performance?What jax/jaxlib version are you using?
Which accelerator(s) are you using?
Additional system info
Python 3.9.16 on Linux
NVIDIA GPU info
Beta Was this translation helpful? Give feedback.
All reactions