Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Bug]: Call RWKV6Attention and report an error environment. #74

Open
synbol opened this issue Oct 31, 2024 · 3 comments
Open

[Bug]: Call RWKV6Attention and report an error environment. #74

synbol opened this issue Oct 31, 2024 · 3 comments
Labels
bug Something isn't working

Comments

@synbol
Copy link

synbol commented Oct 31, 2024

Describe the bug

Error message:
python: /project/lib/Analysis/Allocation.cpp:47: std::pair<llvm::SmallVector, llvm::SmallVector > mlir::triton::getCvtOrder(mlir::Attribute, mlir::Attribute): Assertion `!(srcMmaLayout && dstMmaLayout && !srcMmaLayout.isAmpere()) && "mma -> mma layout conversion is only supported on Ampere"' failed.

Steps to reproduce the bug

Calling process:
from fla.layers.rwkv6 import RWKV6Attention
self.attention = RWKV6Attention(hidden_size=config.dim, num_heads=config.n_head)

o, _, past_key_values = self.attention(self.attention_norm(x), attention_mask=mask, past_key_values=past_key_values)

Expected behavior

None

Environment info

Environment:
torch 2.4.1
triton 3.0.0
einops 0.8.0

@synbol synbol added the bug Something isn't working label Oct 31, 2024
@synbol
Copy link
Author

synbol commented Oct 31, 2024

self.attention = RWKV6Attention(hidden_size=config.dim, num_heads=config.n_head, layer_idx=layer_id)

@yzhangcs
Copy link
Collaborator

Hi, can you provide some runnable code for reproduction. It works normally for me by running this

python benchmark_training_throughput.py --name rwkv6

@sustcsonglin
Copy link
Owner

Hi, thanks for reporting it. What is your GPU model?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working
Projects
None yet
Development

No branches or pull requests

3 participants