Skip to content

v0.2.3

Compare
Choose a tag to compare
@fedebotu fedebotu released this 19 Sep 10:22
· 732 commits to main since this release

Add FlashAttention2 support ⚡

  • Add FlashAttention2 support as mentioned here
  • Remove old wrapper for half() precision since Lightning already deals with this
  • Fix scaled_dot_product_attention implementation in PyTorch < 2.0
  • Minor fixes