|
4bf2ebf836
|
Use u8 tensors for masks. (#273)
|
2023-07-29 11:32:58 +01:00 |
|
|
7513a5e005
|
Line-up the llama implementation with the python-transformers one. (#271)
* Line-up the llama implementation with the python-transformers one.
* Also lineup the multiprocess version.
|
2023-07-28 18:31:28 +01:00 |
|
|
3eb2bc6d07
|
Softmax numerical stability. (#267)
* Softmax numerical stability.
* Fix the flash-attn test.
|
2023-07-28 13:13:01 +01:00 |
|
|
25a2086e8f
|
Putting back Send + Sync
|
2023-07-27 09:58:47 +02:00 |
|
|
7c7e6ba201
|
Removing inner dependency on safetensors.
|
2023-07-27 09:58:47 +02:00 |
|
|
ed58de7551
|
Fixed TP sharded version.
|
2023-07-27 09:58:46 +02:00 |
|
|
1735e4831e
|
TP sharding v2
|
2023-07-27 09:58:14 +02:00 |
|