|
c753f72c85
|
Support for attention bias in gemma + refactor things a bit. (#1744)
* Support for attention bias in gemma + refactor things a bit.
* Fix the cuda tests.
|
2024-02-22 09:35:28 +01:00 |
|
|
63944714f2
|
Use candle_nn::embedding instead of local copies in a few models. (#1562)
|
2024-01-10 21:36:27 +01:00 |
|
|
783735cf22
|
Use softmax-last-dim where possible. (#1057)
|
2023-10-08 13:16:42 +01:00 |
|
|
d3f05eae8c
|
Move some models to candle-transformers so that it's easier to re-use. (#794)
* Move some models to candle-transformers so that they can be shared.
* Also move falcon.
* Move Llama.
* Move whisper (partial).
|
2023-09-10 09:40:27 +01:00 |
|