We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Flash attention integration tests failed
Flash attention integration correctness pytest should pass.
FAILED flash_attention_correctness.py::test_attention[True-dtype0] - TypeError: multiple values for argument 'softmax_scale' FAILED flash_attention_correctness.py::test_attention[True-dtype1] - TypeError: multiple values for argument 'softmax_scale' FAILED flash_attention_correctness.py::test_attention[False-dtype0] - TypeError: multiple values for argument 'softmax_scale' FAILED flash_attention_correctness.py::test_attention[False-dtype1] - TypeError: multiple values for argument 'softmax_scale'
Running
pytest flash_attention_correctness.py
from test/integration/flash_attention.
test/integration/flash_attention
It is currently calling compiler NKI kernel. Upgrade to the one from the repo?
No response
2.21
The text was updated successfully, but these errors were encountered:
No branches or pull requests
Describe the bug
Flash attention integration tests failed
Expected Behavior
Flash attention integration correctness pytest should pass.
Current Behavior
Reproduction Steps
Running
from
test/integration/flash_attention
.Regression Issue
Possible Solution
It is currently calling compiler NKI kernel. Upgrade to the one from the repo?
Additional Information/Context
No response
neuronx-cc version used
2.21
Framework(s) and their versions used (JAX, PyTorch, etc..)
No response
The text was updated successfully, but these errors were encountered: