This website requires JavaScript.
Explore
Help
Register
Sign In
huggingface
/
candle
Watch
1
Star
0
Fork
0
You've already forked candle
mirror of
https://github.com/huggingface/candle.git
synced
2025-06-16 02:38:10 +00:00
Code
Issues
Packages
Projects
Releases
Wiki
Activity
Files
d2c3f1477397b6730fbef7225dd9e5fc0a9fa096
candle
/
candle-flash-attn
History
Laurent Mazare
d2c3f14773
Fix for flash-attn. (
#1310
)
...
Co-authored-by: laurent <
laurent@par2dc5-ai-prd-cl01dgx02.cm.cluster
>
2023-11-10 10:27:27 +01:00
..
cutlass
@
c4f6b8c6bc
Add flash attention (
#241
)
2023-07-26 07:48:10 +01:00
kernels
Add back the bf16 flash-attn kernels. (
#730
)
2023-09-04 07:50:52 +01:00
src
Fix for flash-attn. (
#1310
)
2023-11-10 10:27:27 +01:00
tests
Flash attention without padding (varlen). (
#281
)
2023-07-31 09:45:39 +01:00
build.rs
feat: parse Cuda compute cap from env (
#1066
)
2023-10-16 15:37:38 +01:00
Cargo.toml
Bump the version to 0.3.0. (
#1014
)
2023-10-01 13:51:57 +01:00
README.md
Add some missing readme files. (
#304
)
2023-08-02 10:57:12 +01:00
README.md
candle-flash-attn