You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: llama_cpp/server/app.py
+4-3Lines changed: 4 additions & 3 deletions
Original file line number
Diff line number
Diff line change
@@ -166,10 +166,10 @@ def get_llama():
166
166
description="Positive values penalize new tokens based on their existing frequency in the text so far, decreasing the model's likelihood to repeat the same line verbatim.",
167
167
)
168
168
169
+
169
170
classCreateCompletionRequest(BaseModel):
170
171
prompt: Union[str, List[str]] =Field(
171
-
default="",
172
-
description="The prompt to generate completions for."
172
+
default="", description="The prompt to generate completions for."
0 commit comments