Skip to content

Commit 8d34dcd

Browse files
committed
updated
1 parent 5ff2c14 commit 8d34dcd

File tree

3 files changed

+14
-11
lines changed

3 files changed

+14
-11
lines changed

README.md

+1-2
Original file line numberDiff line numberDiff line change
@@ -17,7 +17,6 @@ from llm_api_utils import get_llm_response
1717

1818
model_name = 'gpt-4o-mini-2024-07-18'
1919
params = {
20-
'max_tokens': 256,
2120
'temperature': 0.0
2221
}
2322

@@ -56,7 +55,7 @@ print('#######', model_name)
5655
[Sample code](./call_gpt.py)
5756
- document: https://platform.openai.com/docs/overview
5857
- library: https://github.com/openai/openai-python
59-
- models: https://platform.openai.com/docs/models
58+
- models: https://platform.openaxi.com/docs/models
6059
- playground: https://platform.openai.com/playground/chat?models=gpt-4o
6160
- pricing: https://openai.com/api/pricing/
6261
- status: https://status.openai.com/

example.py

-1
Original file line numberDiff line numberDiff line change
@@ -3,7 +3,6 @@
33

44
model_name = 'gpt-4o-mini-2024-07-18'
55
params = {
6-
'max_tokens': 4096,
76
'temperature': 0.0
87
}
98

llm_api_utils.py

+13-8
Original file line numberDiff line numberDiff line change
@@ -13,14 +13,17 @@
1313
import google.api_core.exceptions as google_exceptions
1414

1515

16-
def get_llm_response(model_name: str, params: dict, messages: list[dict]) -> str:
16+
def get_llm_response(model_name: str, params_: dict, messages: list[dict]) -> str:
17+
params = params_.copy()
1718
if model_name in OPENAI_MODEL_NAMES:
1819
if 'max_tokens' in params:
19-
params_copy = params.copy()
20-
params_copy['max_completion_tokens'] = params_copy['max_tokens']
21-
del params_copy['max_tokens']
22-
return get_gpt_respnose(model_name, params_copy, messages)
20+
params['maxcompletion_tokens'] = params['max_tokens']
21+
del params['max_tokens']
22+
return get_gpt_respnose(model_name, params, messages)
23+
return get_gpt_respnose(model_name, params, messages)
2324
elif model_name in ANTHROPIC_MODEL_NAMES:
25+
if 'max_tokens' not in params:
26+
params['max_tokens'] = 8192
2427
return get_claude_response(model_name, params, messages)
2528
elif model_name in GEMINI_MODEL_NAMES:
2629
return get_gemini_response(model_name, params, messages)
@@ -32,9 +35,11 @@ def get_llm_response(model_name: str, params: dict, messages: list[dict]) -> str
3235
async def get_llm_response_async(model_name: str, params: dict, messages: list[dict]) -> str:
3336
if model_name in OPENAI_MODEL_NAMES:
3437
if 'max_tokens' in params:
35-
params['max_completion_tokens'] = params['max_tokens']
36-
del params['max_tokens']
37-
return await get_gpt_respnose_async(model_name, params, messages)
38+
params_copy = params.copy()
39+
params_copy['max_completion_tokens'] = params_copy['max_tokens']
40+
del params_copy['max_tokens']
41+
return get_gpt_respnose(model_name, params_copy, messages)
42+
return get_gpt_respnose(model_name, params, messages)
3843
elif model_name in ANTHROPIC_MODEL_NAMES:
3944
return await get_claude_response_async(model_name, params, messages)
4045
elif model_name in GEMINI_MODEL_NAMES:

0 commit comments

Comments
 (0)