We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Error message: python: /project/lib/Analysis/Allocation.cpp:47: std::pair<llvm::SmallVector, llvm::SmallVector > mlir::triton::getCvtOrder(mlir::Attribute, mlir::Attribute): Assertion `!(srcMmaLayout && dstMmaLayout && !srcMmaLayout.isAmpere()) && "mma -> mma layout conversion is only supported on Ampere"' failed.
Calling process: from fla.layers.rwkv6 import RWKV6Attention self.attention = RWKV6Attention(hidden_size=config.dim, num_heads=config.n_head)
o, _, past_key_values = self.attention(self.attention_norm(x), attention_mask=mask, past_key_values=past_key_values)
None
Environment: torch 2.4.1 triton 3.0.0 einops 0.8.0
The text was updated successfully, but these errors were encountered:
self.attention = RWKV6Attention(hidden_size=config.dim, num_heads=config.n_head, layer_idx=layer_id)
Sorry, something went wrong.
Hi, can you provide some runnable code for reproduction. It works normally for me by running this
python benchmark_training_throughput.py --name rwkv6
Hi, thanks for reporting it. What is your GPU model?
No branches or pull requests
Describe the bug
Error message:
python: /project/lib/Analysis/Allocation.cpp:47: std::pair<llvm::SmallVector, llvm::SmallVector > mlir::triton::getCvtOrder(mlir::Attribute, mlir::Attribute): Assertion `!(srcMmaLayout && dstMmaLayout && !srcMmaLayout.isAmpere()) && "mma -> mma layout conversion is only supported on Ampere"' failed.
Steps to reproduce the bug
Calling process:
from fla.layers.rwkv6 import RWKV6Attention
self.attention = RWKV6Attention(hidden_size=config.dim, num_heads=config.n_head)
o, _, past_key_values = self.attention(self.attention_norm(x), attention_mask=mask, past_key_values=past_key_values)
Expected behavior
None
Environment info
Environment:
torch 2.4.1
triton 3.0.0
einops 0.8.0
The text was updated successfully, but these errors were encountered: