Skip to content

Commit 37e3449

Browse files
committed
fix: Gemma4 max_new_tokens 512 → 2048 — responses were truncating mid-answer
1 parent a5cd191 commit 37e3449

2 files changed

Lines changed: 3 additions & 1 deletion

File tree

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1 @@
1+
# Fix: Gemma 4 response truncation — max_new_tokens 512 → 2048

public/ai-worker-gemma4.js

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -213,7 +213,8 @@ async function generate({ userPrompt, prompt, attachments = [], context, chatHis
213213

214214
await model.generate({
215215
...inputs,
216-
max_new_tokens: options.maxTokens || 512,
216+
max_new_tokens: options.maxTokens || 2048,
217+
217218
do_sample: true,
218219
temperature: options.temperature || 0.7,
219220
top_p: options.topP || 0.9,

0 commit comments

Comments
 (0)