Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Benchmarking tracy capture unresponsive on A100 bots #16157

Closed
antiagainst opened this issue Jan 19, 2024 · 0 comments · Fixed by #18144
Closed

Benchmarking tracy capture unresponsive on A100 bots #16157

antiagainst opened this issue Jan 19, 2024 · 0 comments · Fixed by #18144
Labels
hal/cuda Runtime CUDA HAL backend

Comments

@antiagainst
Copy link
Contributor

#16114 disables automatic tracy capture for CUDA benchmarking due to that it's not progressing. See https://github.com/openxla/iree/actions/runs/7545661666/job/20545113878 for a failed job after running 6 hours.

@antiagainst antiagainst added the hal/cuda Runtime CUDA HAL backend label Jan 19, 2024
hanhanW added a commit that referenced this issue Jan 22, 2024
- Revert
llvm/llvm-project@af78e5d
locally. We need to bump tf-nightly to drop the revert.
#16173
- Disable tracy capture for CUDA benchmarking due to that it's not
progressing. #16157
- Apply a
[fix](a863467)
for SME compilation failures.
- Add fixes for
llvm/llvm-project@0a8e3dd

Closes #16063 because
llvm/llvm-project@974ded9
fixes the issue; the commit is carried in the integrate.

---------

Co-authored-by: Benjamin Maxwell <benjamin.maxwell@arm.com>
Co-authored-by: Lei Zhang <antiagainst@gmail.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
hal/cuda Runtime CUDA HAL backend
Projects
None yet
Development

Successfully merging a pull request may close this issue.

1 participant