Add a cuda kernel for dequantizing q8_0. (#1804)

This commit is contained in:
Laurent Mazare
2024-03-05 09:50:37 +01:00
committed by GitHub
parent 8cc0a183ba
commit bd9ab9bc04
2 changed files with 24 additions and 4 deletions

View File

@ -738,10 +738,6 @@ macro_rules! quantized_matmul {
// stable. https://github.com/rust-lang/rust/issues/29599
($fn_name: ident, $fn_name_cpu: ident, $fn_name_cuda: ident, $fn_name_metal: ident, $dtype: expr) => {
fn $fn_name(device: &Device) -> Result<()> {
if device.is_cuda() {
// TODO Enable Cuda GGML sometime maybe.
return Ok(());
}
test_matmul(device, (1, 3, 4, 256), $dtype)?;
Ok(())
}