Mention the flash-attention restriction in the readme. (#1158)

This commit is contained in:
Laurent Mazare
2023-10-23 10:26:56 +01:00
committed by GitHub
parent a11af79e23
commit 25c3cc4149

View File

@ -50,6 +50,9 @@ cached.
Enabling flash-attention requires both a feature flag, `--feature flash-attn` Enabling flash-attention requires both a feature flag, `--feature flash-attn`
and using the command line flag `--use-flash-attn`. and using the command line flag `--use-flash-attn`.
Note that flash-attention-v2 is only compatible with Ampere, Ada, or Hopper GPUs
(e.g., A100/H100, RTX 3090/4090).
## Image to Image Pipeline ## Image to Image Pipeline
... ...