Skip to content

Commit 8740ddc

Browse files
committed
Only support generating one prompt at a time.
1 parent 8895b90 commit 8740ddc

File tree

1 file changed

+4
-3
lines changed

1 file changed

+4
-3
lines changed

llama_cpp/server/app.py

Lines changed: 4 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -166,10 +166,10 @@ def get_llama():
166166
description="Positive values penalize new tokens based on their existing frequency in the text so far, decreasing the model's likelihood to repeat the same line verbatim.",
167167
)
168168

169+
169170
class CreateCompletionRequest(BaseModel):
170171
prompt: Union[str, List[str]] = Field(
171-
default="",
172-
description="The prompt to generate completions for."
172+
default="", description="The prompt to generate completions for."
173173
)
174174
suffix: Optional[str] = Field(
175175
default=None,
@@ -224,7 +224,8 @@ def create_completion(
224224
request: CreateCompletionRequest, llama: llama_cpp.Llama = Depends(get_llama)
225225
):
226226
if isinstance(request.prompt, list):
227-
request.prompt = "".join(request.prompt)
227+
assert len(request.prompt) <= 1
228+
request.prompt = request.prompt[0] if len(request.prompt) > 0 else ""
228229

229230
completion_or_chunks = llama(
230231
**request.dict(

0 commit comments

Comments
 (0)