mirror of
https://github.com/huggingface/candle.git
synced 2025-06-19 11:56:45 +00:00
Fix docs quantized qwen3 (#2955)
* fixed docs quantized-qwen3 README * fixed docs quantized-qwen2-instruct README
This commit is contained in:
@ -8,4 +8,8 @@
|
|||||||
cargo run --example quantized-qwen2-instruct --release -- --prompt "Write a function to count prime numbers up to N."
|
cargo run --example quantized-qwen2-instruct --release -- --prompt "Write a function to count prime numbers up to N."
|
||||||
```
|
```
|
||||||
|
|
||||||
0.5b, 1.5b, 7b and 72b models are available via `--model` argument.
|
0.5b, 1.5b, 7b and 72b models are available via `--which` argument.
|
||||||
|
|
||||||
|
```bash
|
||||||
|
cargo run --release --example quantized-qwen2-instruct -- --which 0.5b --prompt "Write a function to count prime numbers up to N."
|
||||||
|
```
|
||||||
|
@ -8,4 +8,10 @@
|
|||||||
cargo run --example quantized-qwen3 --release -- --prompt "Write a function to count prime numbers up to N."
|
cargo run --example quantized-qwen3 --release -- --prompt "Write a function to count prime numbers up to N."
|
||||||
```
|
```
|
||||||
|
|
||||||
0.6b is used by default, 1.7b, 4b, 8b, 14b, and 32b models are available via `--model` argument.
|
|
||||||
|
0.6b is used by default, 1.7b, 4b, 8b, 14b, and 32b models are available via `--which` argument.
|
||||||
|
|
||||||
|
```bash
|
||||||
|
cargo run --example quantized-qwen3 --release -- --which 4b --prompt "A train is travelling at 120mph, how far does it travel in 3 minutes 30 seconds?"
|
||||||
|
```
|
||||||
|
|
||||||
|
Reference in New Issue
Block a user