Skip to content

Commit 8675397

Browse files
committed
Revert "llama_cpp server: delete some ignored / unused parameters"
This reverts commit b47b954.
1 parent c382d8f commit 8675397

File tree

2 files changed

+16
-3
lines changed

2 files changed

+16
-3
lines changed

llama_cpp/llama_types.py

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -60,6 +60,8 @@ class Completion(TypedDict):
6060
class ChatCompletionMessage(TypedDict):
6161
role: Literal["assistant", "user", "system"]
6262
content: str
63+
user: NotRequired[str]
64+
6365

6466
class ChatCompletionChoice(TypedDict):
6567
index: int

llama_cpp/server/app.py

Lines changed: 14 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -185,7 +185,13 @@ def create_completion(
185185
completion_or_chunks = llama(
186186
**request.dict(
187187
exclude={
188-
"model"
188+
"model",
189+
"n",
190+
"frequency_penalty",
191+
"presence_penalty",
192+
"best_of",
193+
"logit_bias",
194+
"user",
189195
}
190196
)
191197
)
@@ -221,7 +227,7 @@ class Config:
221227
def create_embedding(
222228
request: CreateEmbeddingRequest, llama: llama_cpp.Llama = Depends(get_llama)
223229
):
224-
return llama.create_embedding(**request.dict(exclude={"model"}))
230+
return llama.create_embedding(**request.dict(exclude={"model", "user"}))
225231

226232

227233
class ChatCompletionRequestMessage(BaseModel):
@@ -283,7 +289,12 @@ def create_chat_completion(
283289
completion_or_chunks = llama.create_chat_completion(
284290
**request.dict(
285291
exclude={
286-
"model"
292+
"model",
293+
"n",
294+
"presence_penalty",
295+
"frequency_penalty",
296+
"logit_bias",
297+
"user",
287298
}
288299
),
289300
)

0 commit comments

Comments
 (0)