From 4c72b013c7e48c8baf6a01692a7584b3170e56d3 Mon Sep 17 00:00:00 2001 From: Isotr0py Date: Fri, 7 Feb 2025 01:18:22 +0800 Subject: [PATCH] [Bugfix] Fix unsupported FA version check for Turing GPU (#12828) --- vllm/attention/backends/utils.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/vllm/attention/backends/utils.py b/vllm/attention/backends/utils.py index 3c5028a66d585..e8a34434122c4 100644 --- a/vllm/attention/backends/utils.py +++ b/vllm/attention/backends/utils.py @@ -612,5 +612,5 @@ def flash_attn_version(): return fa_version VLLM_FLASH_ATTN_VERSION = flash_attn_version() -except ImportError: +except (ImportError, AssertionError): VLLM_FLASH_ATTN_VERSION = None