From 4dbe9a28ec868939228b51b64f95e453c4f1e21c Mon Sep 17 00:00:00 2001 From: "dependabot[bot]" <49699333+dependabot[bot]@users.noreply.github.com> Date: Mon, 9 Dec 2024 00:29:02 +0000 Subject: [PATCH] Bump flash-attn from 2.6.3 to 2.7.2.post1 Bumps [flash-attn](https://github.com/Dao-AILab/flash-attention) from 2.6.3 to 2.7.2.post1. - [Release notes](https://github.com/Dao-AILab/flash-attention/releases) - [Commits](https://github.com/Dao-AILab/flash-attention/compare/v2.6.3...v2.7.2.post1) --- updated-dependencies: - dependency-name: flash-attn dependency-type: direct:development update-type: version-update:semver-minor ... Signed-off-by: dependabot[bot] --- setup.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/setup.py b/setup.py index 57acaddf5a..93227fd97b 100644 --- a/setup.py +++ b/setup.py @@ -104,7 +104,7 @@ # Flash 2 group kept for backwards compatibility extra_deps['gpu-flash2'] = [ - 'flash-attn==2.6.3', + 'flash-attn==2.7.2.post1', ] extra_deps['gpu'] = copy.deepcopy(extra_deps['gpu-flash2'])