Replies: 1 comment
-
I would recommend to use CUDA at least 11.6, and we don't support < 11.3. If you can make it work, we would gladly accept a PR, but otherwise it's not something we will fix. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
I mean I want to use block sparse attention which uses triton which requires cuda 11.4. And maybe there is a previous version of xformer using triton that requires cuda of a smaller version? I need cuda 11.0 - 11.1 (max 11.2) due to server restrictions.
Now i have
RuntimeError: Triton requires CUDA 11.4+
with cuda 11.0.Beta Was this translation helpful? Give feedback.
All reactions