From 25c3cc4149304a4f6eec93b2f88aa9c241f8f696 Mon Sep 17 00:00:00 2001 From: Laurent Mazare Date: Mon, 23 Oct 2023 10:26:56 +0100 Subject: [PATCH] Mention the flash-attention restriction in the readme. (#1158) --- candle-examples/examples/stable-diffusion/README.md | 3 +++ 1 file changed, 3 insertions(+) diff --git a/candle-examples/examples/stable-diffusion/README.md b/candle-examples/examples/stable-diffusion/README.md index ee83b3f9..b8736a2a 100644 --- a/candle-examples/examples/stable-diffusion/README.md +++ b/candle-examples/examples/stable-diffusion/README.md @@ -50,6 +50,9 @@ cached. Enabling flash-attention requires both a feature flag, `--feature flash-attn` and using the command line flag `--use-flash-attn`. +Note that flash-attention-v2 is only compatible with Ampere, Ada, or Hopper GPUs +(e.g., A100/H100, RTX 3090/4090). + ## Image to Image Pipeline ...