Skip to content

Commit 8f1b44a

Browse files
authored
Update _posts/2025-10-28-Kimi-K2-Accuracy.md
1 parent f334e9e commit 8f1b44a

File tree

1 file changed

+1
-0
lines changed

1 file changed

+1
-0
lines changed

_posts/2025-10-28-Kimi-K2-Accuracy.md

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -47,6 +47,7 @@ To isolate the problem, I devised a crucial experiment. Instead of using vLLM's
4747
A deeper look revealed that the Kimi tokenizer's `apply_chat_template` function signature includes `**kwargs` to accept extra, model-specific parameters. One such parameter, `add_generation_prompt=True`, is essential for correctly formatting the prompt to signal the start of the assistant's turn, guiding it towards generating a tool call.
4848

4949
A correct prompt should end with special tokens that prime the model to act as the assistant:
50+
5051
```
5152
Correct Prompt Suffix: ...<|im_assistant|>assistant<|im_middle|>
5253
```

0 commit comments

Comments
 (0)