-
Notifications
You must be signed in to change notification settings - Fork 25
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
FlexAttention Output Differs from SDPA #62
Comments
Which example is this? |
I run
I'm not sure what you mean by "which example is this"? |
What version of PyTorch are you using? At least locally I am unable to reproduce this |
PyTorch 2.5.0 |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
I run
python examples/benchmark.py
and encounter the following error - the output from flex attention differs from SDPAI use AWS p3.2xlarge (1 V100 GPU) instance with NVIDIA driver version 550.127.05 and CUDA version 12.4.1
The text was updated successfully, but these errors were encountered: