diff --git a/gpt_llm_trainer_v2,_with_GPT_3_5_Fine_Tuning.ipynb b/gpt_llm_trainer_v2,_with_GPT_3_5_Fine_Tuning.ipynb index 5c17054..6a99c97 100644 --- a/gpt_llm_trainer_v2,_with_GPT_3_5_Fine_Tuning.ipynb +++ b/gpt_llm_trainer_v2,_with_GPT_3_5_Fine_Tuning.ipynb @@ -115,14 +115,14 @@ " \"content\": example\n", " })\n", "\n", - " response = openai.ChatCompletion.create(\n", + " response = openai.chat.completions.create(\n", " model=\"gpt-4\",\n", " messages=messages,\n", " temperature=temperature,\n", " max_tokens=1000,\n", " )\n", "\n", - " return response.choices[0].message['content']\n", + " return response.choices[0].message.content\n", "\n", "# Generate examples\n", "prev_examples = []\n", @@ -153,7 +153,7 @@ "source": [ "def generate_system_message(prompt):\n", "\n", - " response = openai.ChatCompletion.create(\n", + " response = openai.chat.completions.create(\n", " model=\"gpt-4\",\n", " messages=[\n", " {\n", @@ -169,7 +169,7 @@ " max_tokens=500,\n", " )\n", "\n", - " return response.choices[0].message['content']\n", + " return response.choices[0].message.content\n", "\n", "system_message = generate_system_message(prompt)\n", "\n", @@ -348,7 +348,7 @@ { "cell_type": "code", "source": [ - "response = openai.ChatCompletion.create(\n", + "response = openai.chat.completions.create(\n", " model=model_name,\n", " messages=[\n", " {\n", @@ -362,7 +362,7 @@ " ],\n", ")\n", "\n", - "response.choices[0].message['content']" + "response.choices[0].message.content" ], "metadata": { "id": "uxbrmzc5dMuC" @@ -387,4 +387,4 @@ }, "nbformat": 4, "nbformat_minor": 0 -} \ No newline at end of file +}