* GGUF support in the quantized model. * Get the GGUF support to work on llama.
q2k
q3k
q4k
q5k
Minimalist ML framework for Rust