mirror of
https://github.com/huggingface/candle.git
synced 2025-06-17 02:58:50 +00:00

* add quantized version of qwen2 and corresponding example for qwen2-instruct * fix quantized qwen2 clippy error
12 lines
369 B
Markdown
12 lines
369 B
Markdown
# candle-quantized-qwen2-instruct
|
|
|
|
[Qwen2]((https://qwenlm.github.io/blog/qwen2/)) is an upgraded version of Qwen1.5, released by Alibaba Cloud.
|
|
|
|
## Running the example
|
|
|
|
```bash
|
|
cargo run --example quantized-qwen2-instruct --release -- --prompt "Write a function to count prime numbers up to N."
|
|
```
|
|
|
|
0.5b, 1.5b, 7b and 72b models are available via `--model` argument.
|