Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Flash Attention Integration Tests Fail #52

Open
1 task
weiT1993 opened this issue Feb 19, 2025 · 0 comments
Open
1 task

Flash Attention Integration Tests Fail #52

weiT1993 opened this issue Feb 19, 2025 · 0 comments
Labels
bug Something isn't working

Comments

@weiT1993
Copy link

Describe the bug

Flash attention integration tests failed

Expected Behavior

Flash attention integration correctness pytest should pass.

Current Behavior

FAILED flash_attention_correctness.py::test_attention[True-dtype0] - TypeError: multiple values for argument 'softmax_scale'
FAILED flash_attention_correctness.py::test_attention[True-dtype1] - TypeError: multiple values for argument 'softmax_scale'
FAILED flash_attention_correctness.py::test_attention[False-dtype0] - TypeError: multiple values for argument 'softmax_scale'
FAILED flash_attention_correctness.py::test_attention[False-dtype1] - TypeError: multiple values for argument 'softmax_scale'

Reproduction Steps

Running

pytest flash_attention_correctness.py

from test/integration/flash_attention.

Regression Issue

  • Select this option if this issue appears to be a regression.

Possible Solution

It is currently calling compiler NKI kernel. Upgrade to the one from the repo?

Additional Information/Context

No response

neuronx-cc version used

2.21

Framework(s) and their versions used (JAX, PyTorch, etc..)

No response

@weiT1993 weiT1993 added the bug Something isn't working label Feb 19, 2025
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working
Projects
None yet
Development

No branches or pull requests

1 participant