How to include both reasoning_content and content in the conversation history you send back to the model.

#6
by rdsm - opened

Note on thinking-in-context with vLLM: When building multi-turn agentic loops, include both reasoning_content and content in the conversation history you send back to the model. The reasoning content should be re-wrapped in ... tags within the assistant message.

If I am deploying that model on vLLM for public openclaw users , how can I ensure that?

Sign up or log in to comment