|
d7e48234d4
|
Add an erf based gelu op (#900)
* Erf based gelu.
* Add the erf backed gelu.
* Test the new gelu op (which is not gelu_new).
|
2023-09-19 19:54:28 +01:00 |
|
|
6615daf242
|
Tweaks to softmax. (#745)
|
2023-09-05 15:22:27 +01:00 |
|
|
ab36a7f3e3
|
Fix for when f16c is not available. (#614)
|
2023-08-27 07:19:52 +01:00 |
|
|
329f661d9b
|
Trace softmax (#568)
* Trace the softmax op.
* Inline the sum.
* Add min/max vec operations.
|
2023-08-23 15:25:50 +01:00 |
|
|
9a5c7db91a
|
Add support for i64 (#563)
* Add the i64 dtype.
* Adapt the cuda kernels.
|
2023-08-23 10:42:19 +01:00 |
|
|
b8263aa15c
|
Quantized support for f16 and f32 (#457)
* Add f32 as a quantized type.
* Add f16 as a quantized type too.
|
2023-08-15 21:09:37 +01:00 |
|
|
531f23b4d0
|
Rename vec-dot to vec-ops. (#449)
* Rename vec-dot to vec-ops.
* Also bump the crate version.
* Add a currently empty readme.
|
2023-08-15 10:48:57 +01:00 |
|
|
495e0b7580
|
Simd support (#448)
* Import the simd intrinsics in candle-core.
* simd version of reduce-sum.
* Bugfix.
* Fix some clippy lints.
|
2023-08-15 09:50:38 +01:00 |
|