Skip to content

Commit

Permalink
..
Browse files Browse the repository at this point in the history
  • Loading branch information
ShashankMosaicML committed Dec 5, 2024
1 parent 216fcb9 commit 438e0f3
Showing 1 changed file with 1 addition and 1 deletion.
2 changes: 1 addition & 1 deletion tests/models/layers/test_attention.py
Original file line number Diff line number Diff line change
Expand Up @@ -191,7 +191,7 @@ def test_sliding_window(sliding_window_size: int, attn_impl: str):
device = 'cuda'
d = 128
n_heads = 8
seqlen_1 = 8
seqlen_1 = 8 if attn_impl != 'flex' else 128 # FlexAttention requires seqlen to be a multiple of 128 (to compute gradients I think). More info: https://pytorch.org/blog/flexattention/#limitations-and-future-work
bsz = 2

query_1 = torch.randn(bsz, seqlen_1,
Expand Down

0 comments on commit 438e0f3

Please sign in to comment.