Skip to content

Latest commit

 

History

History
166 lines (136 loc) · 8.16 KB

CHANGELOG.md

File metadata and controls

166 lines (136 loc) · 8.16 KB

Changelog

[Main branch]

  • Enable KV parallelism by default
    • No realistic use case will disable KV parallelism, because it virtually kills occupancy in any small-batch/few-head case. Most packages should be using this by default, the same way PyTorch's deterministic mode is disabled by default. Users will still get a warning if PyTorch or NATTEN's deterministic mode is enabled. (#167)
  • Bug fixes
    • Fix rare DDP issue (#167).
    • Fix inconsistencies in docs. (#167)
  • QoL
    • Switch from torch.cuda.amp to torch.amp since the former is deprecated (#168)

[0.17.1] - 2024-05-19

  • Fixed interface for python 3.8 and 3.9

[0.17.0] - 2024-05-02

  • Fused neighborhood attention (FNA) kernels
    • 1D, 2D and 3D Neighborhood Attention are supported,
    • Causal neighborhood attention is implemented,
    • Window (kernel) size, dilation, and causality can be defined per-axis,
    • All GPU architectures since Maxwell (SM50) are supported,
      • SM50 up to SM70 are SIMT-only, but support both FP16 and FP32,
      • SM70 and SM75 target Tensor Cores in FP16, and SIMT-style in FP32,
      • SM80 and above target Tensor Cores in FP16, BF16, and FP32.
    • NATTEN Auto-tuner,
    • Memory preferences and KV parallelism modes,
    • Relative positional biases are only supported in forward pass (inference).
    • Memory layout in FNA is different from existing kernels ([B, *, heads, dim] instead of [B, heads, *, dim].)
      • Eventually this layout can skip over the permute/explicit reshape step in the attention module following the QKV projection.
      • For more refer to Fused vs unfused NA.
  • Naive kernels now implement and allow causal masking,
  • Naive kernels (CPU and CUDA) now allow varying parameters (window size, dilation, causal) across axes,
  • Major bug fix in Volta GEMM kernels
    • The epilogue was different for Volta, and it slipped through unit tests,
    • Tests are now more aggressive, and the issue has been fixed.
  • Memory alignment bug in half RPB gradient kernels fixed

[0.15.1] - 2024-01-24

  • Attention tensors can now be views, which allows combining neighborhood and any other attention pattern (i.e. registers, cross attention tokens, and the like) without extra copies. (#85 and #87).
  • Minor bug fixes (#86 and #94).

[0.15.0] - 2024-01-09

  • Refactored kernels
    • The backend is messy, particularly the CUDA backend. A step in the right direction is at least factoring out duplicated.
    • Out of the 7 operations in NATTEN's backend, 6 have duplicates (really 3 underlying ops with different inputs.)
    • See #26 for more details.
  • 3D Neighborhood Attention: naive CPU and CUDA kernels were added.
  • Major refactoring of the C++ API (#38, #47, #53, and #81)
  • GEMM kernels (#38 and #47)
  • New build system with cmake (#38, #53, #81)
  • Bfloat16 support (#38 and #81)
  • Kepler and Maxwell support (#81)
  • Forward mode automatic differentiation support (#74)
  • Experimental support for Nested Tensors (inference only) (#76)
  • Type checking, clang format, and other typesetting/formatting changes (#80)
  • Added profiling scripts (#81)

[0.14.6] - 2023-03-21

Just a really small update that syncs the changes to the private branch. It's mostly about the changed signature in both QK and AV (both 1D and 2D), where we now take in both kernel size and dilation. Up to 0.14.4 we only took in dilation because QK always took in RPB, and RPB's shape is a function of kernel size, and AV took in attention weights, whose last axis is of size kernel size (kernel size squared in 2D). As of 0.14.5, we support optional RPB, which means now we have to take in kernel size in QK. To make the signatures consistent, we added kernel size to AV as well, along with assertions that kernel sizes match up, but that was not synced from the private branch when we pushed out 0.14.5 because we wanted to support PyTorch 2.0 as soon as possible. 0.14.6 is just mostly to keep the master branch consistent with the latest release, since the signature difference would create an inconsistency between the pip package and the master branch.

[0.14.5] - 2023-03-16

Added

  • Torch 2.0 support
  • Optional RPB.

[0.14.4] - 2022-10-31

Added

  • Python 3.10 and 3.11 wheels!
    • Only for supported torch versions.
  • Support torch 1.13.
  • Tiled NA2D for 3x3 kernels.

Changed

  • Minor changes to the setup script to fix pip install natten.

[0.14.2] - 2022-10-15

Added

  • CPU support!
    • CPP backend for CPU computation.
    • CPU-only builds now supported.
    • Note we only have naive kernels for CPU at the moment. Feel free to open a PR!

Changed

  • Refactored the CPP/CUDA backend.
  • Unit tests for NA1D and NA2D
    • Gradcheck tests in slow and fast mode
    • Gradcheck tests for CPU backend
    • Allclose tests between CPU and CUDA outputs and gradients

[0.14.1] - 2022-10-08

Added

  • NATTEN is now available through PyPI and comes with pre-compiled wheels.

Changed

  • Refactored the CPP/CUDA backend.
  • Dropped LegacyNeighborhoodAttention (pure torch implementation) in favor of the upcoming CPU implementation.

[0.13] - 2022-09-29

Added

  • Added dilation support to all kernels, including tiled NA kernels.

Changed

  • Renamed NeighborhoodAttention to NeighborhoodAttention2D in the python interface.
    • NeighborhoodAttention is now deprecated and will be removed in future versions.
  • Renamed NeighborhoodAttention to NeighborhoodAttention2D in the python interface.
    • NeighborhoodAttention is now deprecated and will be removed in future versions.

[0.12] - 2022-07-09

Added

  • Fixed the race condition in K-Backward and V-Backward kernels.
    • This was handled previously with atomic adds (non-deterministic).
    • Now the kernels compute inverse neighbors and compute gradients without a race condition.
  • "Tiled" Neighborhood Attention kernels for 5x5, 7x7, 9x9, 11x11, and 13x13 window sizes.
    • Applies only to QKRPB-Forward and A-Backward.
    • Only supports dim per head = 32 for now.
      • Try to keep your channels a multiple of 32.
  • Improved FP16 support.
    • Specific kernels for FP16 that use half2 addressing.
    • Different threads per block for FP16.
  • New 1D NA kernels
    • Slightly more efficient.
    • With FP16 support.
    • Arbitrary kernel sizes supported (there's no upper bound of 13x13 like the 2D version).

Changed

  • Window size templating, and some other common bits are factored out into a commons header file.
    • Makes extending support and editing easier.
  • Gradchecks now run in fast mode (faster and less memory usage) unless instructed otherwise.

[0.11a] - 2022-05-12

Added

  • 1D Neighborhood Attention

[0.11] - 2022-04-30

Changed

  • Classification and downstream kernels combined.
  • Refactored cuda extension to natten/.
  • Minor speed improvements.

[0.10]

Initial public release.