|
198009453a
|
Matmul (no batch, no strided, f32, f32 only) sort of done.
|
2023-11-01 17:36:51 +01:00 |
|
|
492d164235
|
More scaffolding, now need to implement matmul (for precompute_cos_sin to work).
|
2023-11-01 16:54:09 +01:00 |
|
|
2d84c16fed
|
First pass (Quantized scaffolding work done + quantized example scaffolding).
|
2023-11-01 15:10:11 +01:00 |
|
|
185b54a33b
|
Make some model cloneable. (#1125)
|
2023-10-18 19:30:47 +01:00 |
|
|
089fc3b584
|
Improve the quantized whisper setup. (#1018)
* Improve the quantized whisper setup.
* Fix the config file paths.
* Use the standard matmul where possible.
|
2023-10-02 17:17:46 +01:00 |
|
|
0a647875ec
|
Use softmax-last-dim in the quantized example. (#848)
|
2023-09-14 17:29:24 +01:00 |
|
|
35f72514f5
|
Move more models to candle-transformers (#796)
* Move dinov2.
* Move efficientnet.
* Move the quantized llama model.
* Move segment-anything.
|
2023-09-10 10:20:18 +01:00 |
|