mirror of
https://github.com/huggingface/candle.git
synced 2025-06-16 10:38:54 +00:00
Mention the flash-attention restriction in the readme. (#1158)
This commit is contained in:
@ -50,6 +50,9 @@ cached.
|
||||
Enabling flash-attention requires both a feature flag, `--feature flash-attn`
|
||||
and using the command line flag `--use-flash-attn`.
|
||||
|
||||
Note that flash-attention-v2 is only compatible with Ampere, Ada, or Hopper GPUs
|
||||
(e.g., A100/H100, RTX 3090/4090).
|
||||
|
||||
## Image to Image Pipeline
|
||||
...
|
||||
|
||||
|
Reference in New Issue
Block a user