|
ee3d290f8b
|
Cuda support for dtype conversions.
|
2023-06-27 09:15:46 +01:00 |
|
|
59a59f41a6
|
Add the cuda mode to llama.
|
2023-06-26 10:06:44 +01:00 |
|
|
d867155ef2
|
Load the weights for llama.
|
2023-06-26 07:23:59 +01:00 |
|
|
7a3101f15f
|
Llama bugfix.
|
2023-06-26 07:07:56 +01:00 |
|
|
97424289d1
|
Fix the llama causal mask inversion.
|
2023-06-25 21:16:54 +01:00 |
|
|
117f014b55
|
Add where_cond and properly apply the causal mask.
|
2023-06-25 21:08:03 +01:00 |
|
|
25bcad290e
|
Fix the causal mask computation.
|
2023-06-25 20:19:30 +01:00 |
|
|
8e404eb125
|
Get a some first inference to work on llama.
|
2023-06-25 18:26:15 +01:00 |
|
|
87c5aab005
|
More llama fixes.
|
2023-06-25 18:08:41 +01:00 |
|
|
60a5598c8b
|
Fix some shape errors.
|
2023-06-25 17:56:59 +01:00 |
|
|
817e4b5005
|
Rework the embeddings so that it works on non-contiguous weights + factor out some code.
|
2023-06-25 17:37:47 +01:00 |
|
|
334524e2c4
|
Take as input slices of tensors as well as slices of &Tensors.
|
2023-06-25 17:07:09 +01:00 |
|
|
90c140ff4b
|
Start sketching the llama example.
|
2023-06-25 13:51:20 +01:00 |
|