Skip to content

Releases: Dao-AILab/flash-attention

fa4-v4.0.0.beta8

08 Apr 08:32
15270e6

Choose a tag to compare

fa4-v4.0.0.beta8 Pre-release
Pre-release

What's Changed

  • fix noisy logger by @drisspg in #2414
  • [AMD ROCm] Fix NaN in FMHA BWD when seq_q=0 by @rocking5566 in #2421
  • Add FA4 CI: GitHub Actions workflow with Apptainer on B200 runner by @Johnsonms in #2393
  • Fix some bugs of CI by @Johnsonms in #2423
  • [ROCM] Fix windows issues by @micmelesse in #2385
  • fix: add [cu13] extra to dev install instructions for CUDA 13 / B200 systems by @Johnsonms in #2430
  • Fix: disable 2-CTA backward mode when block_sparse_tensors is used by @jduprat in #2433
  • CI: extend FA4 test matrix with causal/non-causal correctness and fwd+bwd benchmark seqlen 1K-32K by @Johnsonms in #2428

Full Changelog: fa4-v4.0.0.beta7...fa4-v4.0.0.beta8

fa4-v4.0.0.beta7

01 Apr 08:35
f6a16e1

Choose a tag to compare

fa4-v4.0.0.beta7 Pre-release
Pre-release

What's Changed

New Contributors

Full Changelog: fa4-v4.0.0.beta5...fa4-v4.0.0.beta7

fa4-v4.0.0.beta6

25 Mar 08:21
6362bd3

Choose a tag to compare

fa4-v4.0.0.beta6 Pre-release
Pre-release

What's Changed

New Contributors

Full Changelog: fa4-v4.0.0.beta4...fa4-v4.0.0.beta6

fa4-v4.0.0.beta5

23 Mar 16:50
6362bd3

Choose a tag to compare

fa4-v4.0.0.beta5 Pre-release
Pre-release

What's Changed

New Contributors

Full Changelog: fa4-v4.0.0.beta4...fa4-v4.0.0.beta5

fa4-v4.0.0.beta4

05 Mar 18:02

Choose a tag to compare

fa4-v4.0.0.beta2

05 Mar 12:22

Choose a tag to compare

fa4-v4.0.0.beta1

05 Mar 12:19

Choose a tag to compare

fa4-v4.0.0.beta0

05 Mar 12:09

Choose a tag to compare

What's Changed

Read more

v2.8.3

14 Aug 17:12

Choose a tag to compare

Bump to v2.8.3

v2.8.2

24 Jul 05:45

Choose a tag to compare

Bump to v2.8.2