Implement Flash Attention 2 for webgpu EP #23576
+525
−0
Merged
Azure Pipelines / Big Models (Llama2_7B_ONNX Llama2_7B_ONNX)
succeeded
Feb 6, 2025 in 1h 12m 7s
Llama2_7B_ONNX Llama2_7B_ONNX succeeded
Loading