QUESTION: It seems that LiteLLM does not handle the reasoning_content of the model. #3416
Replies: 3 comments
-
|
qwen and deepseek not suport outputting reasoning_content,Even though Litellm supports Deepseek output reasoning_content。 |
Beta Was this translation helpful? Give feedback.
-
|
代码里都没有处理reasoning_content的逻辑 |
Beta Was this translation helpful? Give feedback.
-
|
Response from ADK Answering Agent (experimental, answer may be inaccurate) TLDR: The Hello! Thanks for your question. The behavior you're observing is expected. The When you use a model through The To fix this, you should switch your planner from from adk.agents.planners import PlanReActPlanner
# ... your existing code ...
# Replace BuiltInPlanner with PlanReActPlanner
# base_planner = BuiltInPlanner(thinking_config=ThinkingConfig(include_thoughts=True))
base_planner = PlanReActPlanner(thinking_config=ThinkingConfig(include_thoughts=True))
agent = Agent(
name="simple_agent",
model=llm,
instruction="你是一个乐于助不的中文助手",
description="回答用户的python编程问题",
planner=base_planner
)
# ... rest of your code ...By making this change, you should be able to see the model's reasoning and thought process in the events generated by the runner. This response was generated by an automated assistant. Please note that the information provided may not be exhaustive. |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
When I was connecting to the model via LiteLLM, I noticed that it doesn't seem to support outputting the model's reasoning_content. Here's the content of my code:
I saw the following code in the source code parsing. It seems that LiteLLM does not wrap the reasoning_content output by the model. Is there a problem with my usage or is it indeed not supported?
Beta Was this translation helpful? Give feedback.
All reactions