[Bug] CUDA error: an illegal memory access was encountered on v.0.0.18
#406
Replies: 2 comments 1 reply
-
I'll try to see if I can get access to an L4 VM and maybe reproduce this, but I am a little swamped. Also a little confused if 0.0.17 works, since there shouldn't really be any changes to inference for 0.0.18. |
Beta Was this translation helpful? Give feedback.
1 reply
-
Unclear why, but reinstalling today worked.. Very weird. Couldn't find why. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Hello, I have encountered this new CUDA error on the latest tag. I initially transitioned from
v0.0.12
(which worked) tov0.0.18
(which got this error).v0.0.17
from source worked.My env is a GCP L4 VM, cuda 11.8. I've made sure to use
torch==2.2.0
and clear torch extensions cache. I also tried changing VM gpus, updating dependencies, installing from wheels vs source.Full stack:
I do not see this function
safe_move_tensor
being changed itself. I tried thestreaming.py
example, and it usually runs, but it also sometimes throw this error.The code that generates this issue can be trimmed to:
Beta Was this translation helpful? Give feedback.
All reactions