-
Notifications
You must be signed in to change notification settings - Fork 308
Issues: NVIDIA/TransformerEngine
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
Passing nonexistent argument when flash_attn version is >= 2.5.7
#1213
opened Sep 27, 2024 by
MaciejBalaNV
No option to change FP8 status in graphed module after using "make_graphed_callables"
bug
Something isn't working
#1207
opened Sep 26, 2024 by
MaciejBalaNV
[PyTorch] fused CUDNN attention kernel and sliding window attention
#1197
opened Sep 23, 2024 by
Marks101
[PyTorch] fused CUDNN attention kernel not properly handling strides
#1195
opened Sep 23, 2024 by
Marks101
[PyTorch] FP8 and activation checkpointing causes training instabilities
#1190
opened Sep 18, 2024 by
Marks101
AssertionError: Outputs not close enough in tensor in test_numerics.py
bug
Something isn't working
#1165
opened Sep 6, 2024 by
sirutBuasai
AssertionError: Device compute capability 8.9 or higher required for FP8 execution.
#1159
opened Sep 5, 2024 by
kamrul-NSL
[PyTorch] Bug in FP8 buffer update causing training instabilities
#1047
opened Jul 26, 2024 by
Marks101
Previous Next
ProTip!
Type g i on any issue or pull request to go back to the issue listing page.