13
13
import google .api_core .exceptions as google_exceptions
14
14
15
15
16
- def get_llm_response (model_name : str , params : dict , messages : list [dict ]) -> str :
16
+ def get_llm_response (model_name : str , params_ : dict , messages : list [dict ]) -> str :
17
+ params = params_ .copy ()
17
18
if model_name in OPENAI_MODEL_NAMES :
18
19
if 'max_tokens' in params :
19
- params_copy = params . copy ()
20
- params_copy [ 'max_completion_tokens' ] = params_copy ['max_tokens' ]
21
- del params_copy [ 'max_tokens' ]
22
- return get_gpt_respnose (model_name , params_copy , messages )
20
+ params [ 'maxcompletion_tokens' ] = params [ 'max_tokens' ]
21
+ del params ['max_tokens' ]
22
+ return get_gpt_respnose ( model_name , params , messages )
23
+ return get_gpt_respnose (model_name , params , messages )
23
24
elif model_name in ANTHROPIC_MODEL_NAMES :
25
+ if 'max_tokens' not in params :
26
+ params ['max_tokens' ] = 8192
24
27
return get_claude_response (model_name , params , messages )
25
28
elif model_name in GEMINI_MODEL_NAMES :
26
29
return get_gemini_response (model_name , params , messages )
@@ -32,9 +35,11 @@ def get_llm_response(model_name: str, params: dict, messages: list[dict]) -> str
32
35
async def get_llm_response_async (model_name : str , params : dict , messages : list [dict ]) -> str :
33
36
if model_name in OPENAI_MODEL_NAMES :
34
37
if 'max_tokens' in params :
35
- params ['max_completion_tokens' ] = params ['max_tokens' ]
36
- del params ['max_tokens' ]
37
- return await get_gpt_respnose_async (model_name , params , messages )
38
+ params_copy = params .copy ()
39
+ params_copy ['max_completion_tokens' ] = params_copy ['max_tokens' ]
40
+ del params_copy ['max_tokens' ]
41
+ return get_gpt_respnose (model_name , params_copy , messages )
42
+ return get_gpt_respnose (model_name , params , messages )
38
43
elif model_name in ANTHROPIC_MODEL_NAMES :
39
44
return await get_claude_response_async (model_name , params , messages )
40
45
elif model_name in GEMINI_MODEL_NAMES :
0 commit comments