Releases: bdashore3/flash-attention
Releases Β· bdashore3/flash-attention
v2.7.1.post1
Actions: Bump Cuda 12.4 Signed-off-by: kingbri <[email protected]>
v2.7.0.post2
Actions: Bump Cuda 12.4 Signed-off-by: kingbri <[email protected]>
v2.6.3
Synced to Upstream version
NOTE: Backward and Dropout are disabled meaning that this release is INFERENCE ONLY.
This is because including these features more than doubles the build time and makes the github action time itself out. Please raise an issue to the parent repo to help reduce the build times if you want these features.
v2.6.1
Actions: Switch to CUDA 12.3 Signed-off-by: kingbri <[email protected]>
v2.5.9.post2
Quick release to add softcapping commits. Does not have backward, dropout, or alibi support.
v2.5.9.post1
Actions: Clarify dispatch formatting Signed-off-by: kingbri <[email protected]>
v2.5.8
Same as Upstream tag
Now built for only torch 2.2.2 and 2.3.0
v2.5.6
Same as Upstream tag
v2.5.2
Same as the upstream tag
Adds this PR to help fix building on Windows
v2.4.2
Inline with the parent repo's tag
Made for cuda 12.x and pytorch 2.1.2 and 2.2
v2.4.3 and up cannot be built on Windows at this time.