Files
candle/candle-examples/examples
Laurent Mazare 30a958e5dd Quantized mixtral model (#1442)
* Add the Mixtral model.

* Add more of the mixtral layers.

* Add the final layers for mixtral.

* Sketch the expert selection.

* Add some expert routing logic.

* Hopefully finish the routing logic for mixtral.

* Add the mixtral example.

* Fix the weight filenames.

* Bugfix.

* Another fix.

* Yet another fix + remove the unused pragma.

* Shape fix.

* Support for quantized mixtral.

* Support mixtral in the quantized example.

* Mlp or moe type.

* Fix the expert field namings.

* Refactor the mlp bit.

* More MoE logic.

* Add the MoE quantized logic.

* Fix the experts length.
2023-12-15 19:16:06 -06:00
..
2023-10-11 19:51:10 +01:00
2023-11-24 15:09:14 +00:00
2023-12-15 14:19:56 -06:00
2023-12-15 07:01:10 -06:00
2023-12-15 19:16:06 -06:00
2023-11-12 09:46:19 +00:00
2023-10-20 09:08:39 +01:00