Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Run pre-submit on all files. #2

Merged
merged 1 commit into from
Apr 22, 2024
Merged

Run pre-submit on all files. #2

merged 1 commit into from
Apr 22, 2024

Conversation

stellaraccident
Copy link
Collaborator

No description provided.

@stellaraccident stellaraccident merged commit dce7436 into main Apr 22, 2024
3 checks passed
@stellaraccident stellaraccident deleted the run_presubmit_all branch April 22, 2024 01:14
harsh-nod added a commit to harsh-nod/iree-turbine that referenced this pull request Jul 16, 2024
harsh-nod added a commit to harsh-nod/iree-turbine that referenced this pull request Jul 16, 2024
Signed-off-by: Harsh Menon <harsh@nod-labs.com>
harsh-nod added a commit to harsh-nod/iree-turbine that referenced this pull request Aug 27, 2024
Signed-off-by: Harsh Menon <harsh@nod-labs.com>
harsh-nod added a commit to harsh-nod/iree-turbine that referenced this pull request Oct 30, 2024
Signed-off-by: Harsh Menon <harsh@nod-labs.com>
harsh-nod added a commit to harsh-nod/iree-turbine that referenced this pull request Dec 4, 2024
nicolasvasilache added a commit to nicolasvasilache/iree-turbine that referenced this pull request Feb 14, 2025
Avoid flex_attention for RPE as it is unclear there is a correct implementation possible given the
limitations of the create_block_mask re conditionals.

Instead we use the manual torch implementation that is known to be correct.

As we update the test and the extend_attention_rpe to use a static max_rpe_context_length,
a new error appears that suggests some issue with the indexing in extend_attention_rpe.

Repro:
```
pytest tests/kernel/wave/attention/extend_attention_test.py --run-e2e -v -k "rpe"
```

Errors out with:
```
E               Diagnostics:
E               <stdin>:282:18: error: 'vector.gather' op operand iree-org#2 must be vector of integer or index values, but got 'index'
E                         %468 = "vector.gather"(%109, %39, %39, %467, %44) : (memref<?xf32, strided<[1], offset: ?>>, index, index, vector<4xi1>, vector<4xf32>) -> vector<4xf32>
E                                ^
```

Signed-off-by: Nicolas Vasilache <nicolasvasilache@users.noreply.github.com>
nicolasvasilache added a commit to nicolasvasilache/iree-turbine that referenced this pull request Feb 14, 2025
Avoid flex_attention for RPE as it is unclear there is a correct implementation possible given the
limitations of the create_block_mask re conditionals.

Instead we use the manual torch implementation that is known to be correct.

As we update the test and the extend_attention_rpe to use a static max_rpe_context_length,
a new error appears that suggests some issue with the indexing in extend_attention_rpe.

Repro:
```
pytest tests/kernel/wave/attention/extend_attention_test.py --run-e2e -v -k "rpe"
```

Errors out with:
```
E               Diagnostics:
E               <stdin>:282:18: error: 'vector.gather' op operand iree-org#2 must be vector of integer or index values, but got 'index'
E                         %468 = "vector.gather"(%109, %39, %39, %467, %44) : (memref<?xf32, strided<[1], offset: ?>>, index, index, vector<4xi1>, vector<4xf32>) -> vector<4xf32>
E                                ^
```

Signed-off-by: Nicolas Vasilache <nicolasvasilache@users.noreply.github.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

1 participant