-
Notifications
You must be signed in to change notification settings - Fork 4
/
Copy pathexample.py
36 lines (30 loc) · 1.18 KB
/
example.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
from llm_api_utils import get_llm_response
model_name = 'gpt-4o-mini-2024-07-18'
params = {
'temperature': 0.0
}
messages = [
{"role": "system", "content": "回答の際は、3つの回答を箇条書きで回答してください。"},
{"role": "user", "content": "大喜利しましょう。とても面白い回答をしてくださいね。"},
{"role": "assistant", "content": "おけ、任せて"},
{"role": "user", "content": "こんな台風は嫌だ、どんな台風?"}
]
print(messages[-1]['content'])
response = get_llm_response(model_name, params, messages)
print(response)
print('#######', model_name)
model_name = 'claude-3-5-sonnet-20240620'
response = get_llm_response(model_name, params, messages)
print(response)
print('#######', model_name)
# example of asynchronous request
import asyncio
from llm_api_utils import get_llm_response_async
from tenacity import retry, stop_after_attempt, wait_fixed
model_name = 'models/gemini-1.5-pro-exp-0827'
@retry(wait=wait_fixed(90), stop=stop_after_attempt(10))
async def main():
response = await get_llm_response_async(model_name, params, messages)
print(response)
asyncio.run(main())
print('#######', model_name)