mirror of
https://github.com/huggingface/candle.git
synced 2025-06-17 11:08:52 +00:00
fix firstToken, minor ui changes (#971)
This commit is contained in:
@ -50,6 +50,7 @@ async function generate(data) {
|
||||
tokenizerURL,
|
||||
prompt,
|
||||
temp,
|
||||
top_p,
|
||||
repeatPenalty,
|
||||
seed,
|
||||
maxSeqLen,
|
||||
@ -59,11 +60,17 @@ async function generate(data) {
|
||||
const model = await Llama2C.getInstance(weightsURL, modelID, tokenizerURL);
|
||||
|
||||
self.postMessage({ status: "loading", message: "Initializing model" });
|
||||
model.init_with_prompt(prompt, temp, repeatPenalty, seed);
|
||||
const firstToken = model.init_with_prompt(
|
||||
prompt,
|
||||
temp,
|
||||
top_p,
|
||||
repeatPenalty,
|
||||
seed
|
||||
);
|
||||
|
||||
const seq_len = model.get_seq_len();
|
||||
|
||||
let sentence = "";
|
||||
let sentence = firstToken;
|
||||
let maxTokens = maxSeqLen ? maxSeqLen : seq_len - prompt.length - 1;
|
||||
let startTime = performance.now();
|
||||
let tokensCount = 0;
|
||||
|
Reference in New Issue
Block a user