From 89a00b56cce79e1735928701c0e45f5c70122adb Mon Sep 17 00:00:00 2001 From: cksac Date: Wed, 23 Aug 2023 18:05:33 +0800 Subject: [PATCH] add chat models in quantized example (#551) * add chat models in quantized example * cargo fmt --- candle-examples/examples/quantized/main.rs | 18 ++++++++++++++++++ 1 file changed, 18 insertions(+) diff --git a/candle-examples/examples/quantized/main.rs b/candle-examples/examples/quantized/main.rs index 7c457f7a..c4be8974 100644 --- a/candle-examples/examples/quantized/main.rs +++ b/candle-examples/examples/quantized/main.rs @@ -391,6 +391,12 @@ enum Which { L13b, #[value(name = "70b")] L70b, + #[value(name = "7b-chat")] + L7bChat, + #[value(name = "13b-chat")] + L13bChat, + #[value(name = "70b-chat")] + L70bChat, } #[derive(Parser, Debug)] @@ -466,6 +472,18 @@ impl Args { Which::L7b => ("TheBloke/Llama-2-7B-GGML", "llama-2-7b.ggmlv3.q4_0.bin"), Which::L13b => ("TheBloke/Llama-2-13B-GGML", "llama-2-13b.ggmlv3.q4_0.bin"), Which::L70b => ("TheBloke/Llama-2-70B-GGML", "llama-2-70b.ggmlv3.q4_0.bin"), + Which::L7bChat => ( + "TheBloke/Llama-2-7B-Chat-GGML", + "llama-2-7b-chat.ggmlv3.q4_0.bin", + ), + Which::L13bChat => ( + "TheBloke/Llama-2-13B-Chat-GGML", + "llama-2-13b-chat.ggmlv3.q4_0.bin", + ), + Which::L70bChat => ( + "TheBloke/Llama-2-70B-Chat-GGML", + "llama-2-70b-chat.ggmlv3.q4_0.bin", + ), }; let api = hf_hub::api::sync::Api::new()?; let api = api.model(repo.to_string());