We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
test_lm_eval_accuracy_v1_engine[google/gemma-3-1b-it]
1 parent 3b45075 commit 9f020f4Copy full SHA for 9f020f4
cmake/external_projects/vllm_flash_attn.cmake
@@ -38,7 +38,7 @@ else()
38
FetchContent_Declare(
39
vllm-flash-attn
40
GIT_REPOSITORY https://github.com/vllm-project/flash-attention.git
41
- GIT_TAG 8f468e7da54a8e2f98abfa7c38636aac91c0cba1
+ GIT_TAG a893712401d70362fbb299cd9c4b3476e8e9ed54
42
GIT_PROGRESS TRUE
43
# Don't share the vllm-flash-attn build between build types
44
BINARY_DIR ${CMAKE_BINARY_DIR}/vllm-flash-attn
0 commit comments