Skip to content

Commit

Permalink
fix: add back typo removed variable
Browse files Browse the repository at this point in the history
  • Loading branch information
drbh committed Feb 21, 2024
1 parent 3e22cdd commit 66f8912
Showing 1 changed file with 2 additions and 0 deletions.
2 changes: 2 additions & 0 deletions server/text_generation_server/models/__init__.py
Original file line number Diff line number Diff line change
Expand Up @@ -43,6 +43,8 @@

FLASH_ATT_ERROR_MESSAGE = "{} requires Flash Attention enabled models."

FLASH_ATTENTION = True

# FlashCausalLM reqiures CUDA Graphs to be enabled on the system. This will throw a RuntimeError
# if CUDA Graphs are not available when calling `torch.cuda.graph_pool_handle()` in the FlashCausalLM
HAS_CUDA_GRAPH = False
Expand Down

0 comments on commit 66f8912

Please sign in to comment.