mirror of
https://github.com/huggingface/candle.git
synced 2025-06-16 10:38:54 +00:00
Mention the flash-attention restriction in the readme. (#1158)
This commit is contained in:
@ -50,6 +50,9 @@ cached.
|
|||||||
Enabling flash-attention requires both a feature flag, `--feature flash-attn`
|
Enabling flash-attention requires both a feature flag, `--feature flash-attn`
|
||||||
and using the command line flag `--use-flash-attn`.
|
and using the command line flag `--use-flash-attn`.
|
||||||
|
|
||||||
|
Note that flash-attention-v2 is only compatible with Ampere, Ada, or Hopper GPUs
|
||||||
|
(e.g., A100/H100, RTX 3090/4090).
|
||||||
|
|
||||||
## Image to Image Pipeline
|
## Image to Image Pipeline
|
||||||
...
|
...
|
||||||
|
|
||||||
|
Reference in New Issue
Block a user