39157346cb
Add SAM UI Demo ( #854 )
...
* fix tensor flattening
* send image data back
* sam ui worker example
* SAM example
* resize container
* no need for this
2023-09-15 06:31:58 +01:00
5cefbba757
minor UI fixes ( #856 )
...
* fixes
* remove listener
* remove event listener
2023-09-15 06:30:50 +01:00
805bf9ffa7
Implement top_p / nucleus sampling ( #819 )
...
* Implement top_p / nucleus sampling
* Update changelog
* rustfmt
* Add tests
* Fix clippy warning
* Fix another clippy error
2023-09-12 18:10:16 +02:00
2257f4d475
Bump the crate version + update the changelog. ( #822 )
2023-09-12 06:39:24 +01:00
90e077e409
Return the low res mask in the wasm segment-anything module. ( #798 )
...
* Return the low res mask.
* Add some validations.
2023-09-10 13:03:02 +01:00
584171cae1
Add a wasm module for the segment anything example. ( #797 )
2023-09-10 12:29:37 +01:00
16bf44f6e9
force model cache ( #751 )
2023-09-06 05:53:31 +02:00
6a40decc76
Minor WASM UI improvements ( #748 )
...
* add stats
* random seed btn
* minor ui improvoments
2023-09-05 19:24:43 +01:00
0d00c06a83
Fix clippy lint. ( #736 )
2023-09-04 16:09:19 +01:00
8395152d20
Llama2c WASM UI improvements ( #732 )
...
* pass seed, expose model seq_len
* wip new llama2.c ui
* final new UI example
* small coppy
* copy
2023-09-04 15:59:22 +01:00
e2f9f60ac2
Avoid some redundant clone. ( #731 )
2023-09-04 09:18:32 +02:00
2c1df6bba1
Add a repeat penality to the llama2-c command line example. ( #713 )
...
* Add a repeat penality to the llama2-c command line example.
* Another fix attempt.
2023-09-01 20:38:58 +01:00
4d56cef583
Handle the empty sequence case properly. ( #712 )
...
* Handle the empty sequence case properly.
* Proper fix.
2023-09-01 20:12:30 +01:00
731e3ffb03
Remove unnecessary file. ( #710 )
2023-09-01 19:42:23 +01:00
2fef14cb14
Add a repeat penalty to the llama2.c wasm example. ( #709 )
2023-09-01 19:32:28 +01:00
94aa234dfd
Add the kv-cache to the whisper wasm version. ( #689 )
...
* Add the kv-cache to the whisper wasm version.
* Improve the handling of special tokens.
2023-08-31 09:37:44 +01:00
8e84d8a59b
Llama2.c wasm module. ( #686 )
2023-08-31 07:44:32 +01:00
9bd486fb96
Add Yolo Pose to JS Example ( #684 )
...
* add support for yolo pose models
* fix copy
2023-08-31 06:32:57 +01:00
1d0bb48fae
Improve Whisper WASM UI example ( #669 )
...
* wip add module and js worker example
* params
* clean up, send error
* final UI with whisper webworker
* add simple instructions
2023-08-30 20:35:41 +02:00
618f4e4c78
Add some documentation. ( #673 )
...
* Add some documentation.
* Bump the crate version.
2023-08-30 11:54:00 +01:00
2d3fcad267
Simplify usage of the pool functions. ( #662 )
...
* Simplify usage of the pool functions.
* Small tweak.
* Attempt at using apply to simplify the convnet definition.
2023-08-29 19:12:16 +01:00
a044907ffc
Dilated convolutions ( #657 )
...
* Add the dilation parameter.
* Restore the basic optimizer example.
* Dilation support in cudnn.
* Use the dilation parameter in the cpu backend.
* More dilation support.
* No support for dilation in transposed convolutions.
* Add dilation to a test.
* Remove a print.
* Helper function.
2023-08-29 16:12:11 +01:00
24dda44c27
Add wasm support for yolo-v8 pose detection. ( #630 )
...
* Add wasm support for yolo-v8 pose detection.
* Better bbox handling.
* Add the pose model in the wasm example lib.
2023-08-27 19:49:24 +01:00
72ebb12bca
Remove some dead-code annotations. ( #629 )
...
* Remove some dead-code annotations.
* More dead code removal.
* One more.
* CI fix.
2023-08-27 18:52:33 +01:00
a3f97c143d
Bump the crate version + update CHANGELOG. ( #628 )
2023-08-27 18:17:11 +01:00
864227edbf
[WIP] Improve Yolo WASM UI example ( #591 )
...
* return detections with classes names
* ignore .DS_Store
* example how to load wasm module
* add param to set model size
* add param for model size
* accept iou and confidence threshold on run
* conf and iou thresholds
* clamp only
* remove images from branch
* a couple of renamings, add readme with instructions
* final design
* minor font + border update
2023-08-26 11:40:41 +01:00
aba1e90797
Add some group parameter to convolutions. ( #566 )
...
* Add some group parameter to convolutions.
* Avoid some unnecessary groups checks.
* Move the tensor convolution bits.
* Properh handling of groups.
* Bump the crate version.
* And add a changelog.
2023-08-23 12:58:55 +01:00
7687a0f453
Also fix the aspect ratio in the wasm example. ( #556 )
...
* Also fix the aspect ratio in the wasm example.
* Add the yolo lib.
* Update the build script.
2023-08-22 22:20:08 +01:00
20ce3e9f39
Sketch the yolo wasm example. ( #546 )
...
* Sketch the yolo wasm example.
* Web ui.
* Get the web ui to work.
* UI tweaks.
* More UI tweaks.
* Use the natural width/height.
* Add a link to the hf space in the readme.
2023-08-22 11:56:43 +01:00
a8f61e66cc
Bump the crates version to 0.1.2. ( #522 )
2023-08-20 08:07:07 +01:00
c78ce76501
Add a simple Module trait and implement it for the various nn layers ( #500 )
...
* Start adding the module trait.
* Use the module trait.
* Implement module for qmatmul.
2023-08-18 09:38:22 +01:00
13401df4d1
Add an abstract type for RmsNorm. ( #499 )
2023-08-18 08:52:14 +01:00
d32e8199cd
Layer norm tweaks ( #482 )
...
* Add some options to make layer-norm more configurable.
* Add the rms-norm variant.
* Replace the RmsNorm with the shared bits.
2023-08-17 10:07:13 +01:00
531f23b4d0
Rename vec-dot to vec-ops. ( #449 )
...
* Rename vec-dot to vec-ops.
* Also bump the crate version.
* Add a currently empty readme.
2023-08-15 10:48:57 +01:00
c84883ecf2
Add a cuda kernel for upsampling. ( #441 )
...
* Add a cuda kernel for upsampling.
* Update for the latest tokenizers version.
2023-08-14 13:12:17 +01:00
52414ba5c8
Bugfix for the llama2 wasm example. ( #310 )
...
* Clean-up the llama2.c wasm example.
* Use a proper tokenizer.
* Add a prompt.
* Bugfix for the llama2 wasm example.
2023-08-02 17:32:36 +01:00
186c308d51
Wasm llama2 tweaks ( #309 )
...
* Clean-up the llama2.c wasm example.
* Use a proper tokenizer.
2023-08-02 15:49:43 +01:00
4fe8a02f88
Update the repo location. ( #305 )
2023-08-02 11:12:18 +01:00
d38943aadc
Add version numbers for all the candle crates ( #303 )
...
* Switch to candle-gemm for the time being.
* Add the missing versions.
2023-08-02 10:52:13 +01:00
51e51da896
Rename the candle crate to candle-core ( #301 )
...
* Rename to candle-core.
* More candle-core renaming.
2023-08-02 08:20:22 +01:00
ba2254556c
Display the temperature being used for text generation. ( #278 )
2023-07-30 09:53:05 +01:00
4bf2ebf836
Use u8 tensors for masks. ( #273 )
2023-07-29 11:32:58 +01:00
3eb2bc6d07
Softmax numerical stability. ( #267 )
...
* Softmax numerical stability.
* Fix the flash-attn test.
2023-07-28 13:13:01 +01:00
7c7e6ba201
Removing inner dependency on safetensors.
2023-07-27 09:58:47 +02:00
1735e4831e
TP sharding v2
2023-07-27 09:58:14 +02:00
209f06d7c3
Micro-cleanup. ( #256 )
2023-07-27 07:55:54 +01:00
81bfa46702
Updated.
2023-07-26 15:21:50 +02:00
035372248e
Simple QOL.
...
- Add ms/token on llama2.c (15ms/token on my personal machine)
- Hide `Run` buttons while models are not ready
- Add dummy `progress` while weights are downloading (I briefly looked
at putting a real progressbar.. and nothing easy enough came up.)
2023-07-26 15:17:32 +02:00
97990f4afc
Add number of tokens.
2023-07-26 14:57:20 +02:00
160ba09d30
Polish the llama2 wasm ui. ( #232 )
...
* Polish the llama2 wasm ui.
* readme update.
2023-07-24 15:28:27 +01:00