@@ -83,13 +83,7 @@ class CreateCompletionRequest(BaseModel):
83
83
# ignored, but marked as required for the sake of compatibility with openai's api
84
84
model : str = model_field
85
85
86
- n : Optional [int ] = 1
87
86
logprobs : Optional [int ] = Field (None )
88
- presence_penalty : Optional [float ] = 0
89
- frequency_penalty : Optional [float ] = 0
90
- best_of : Optional [int ] = 1
91
- logit_bias : Optional [Dict [str , float ]] = Field (None )
92
- user : Optional [str ] = Field (None )
93
87
94
88
# llama.cpp specific parameters
95
89
top_k : int = 40
@@ -120,13 +114,7 @@ def create_completion(
120
114
completion_or_chunks = llama (
121
115
** request .dict (
122
116
exclude = {
123
- "model" ,
124
- "n" ,
125
- "frequency_penalty" ,
126
- "presence_penalty" ,
127
- "best_of" ,
128
- "logit_bias" ,
129
- "user" ,
117
+ "model"
130
118
}
131
119
)
132
120
)
@@ -141,7 +129,6 @@ class CreateEmbeddingRequest(BaseModel):
141
129
# ignored, but marked as required for the sake of compatibility with openai's api
142
130
model : str = model_field
143
131
input : str
144
- user : Optional [str ]
145
132
146
133
class Config :
147
134
schema_extra = {
@@ -161,7 +148,7 @@ class Config:
161
148
def create_embedding (
162
149
request : CreateEmbeddingRequest , llama : llama_cpp .Llama = Depends (get_llama )
163
150
):
164
- return llama .create_embedding (** request .dict (exclude = {"model" , "user" }))
151
+ return llama .create_embedding (** request .dict (exclude = {"model" }))
165
152
166
153
167
154
class ChatCompletionRequestMessage (BaseModel ):
@@ -181,12 +168,6 @@ class CreateChatCompletionRequest(BaseModel):
181
168
182
169
# ignored, but marked as required for the sake of compatibility with openai's api
183
170
model : str = model_field
184
-
185
- n : Optional [int ] = 1
186
- presence_penalty : Optional [float ] = 0
187
- frequency_penalty : Optional [float ] = 0
188
- logit_bias : Optional [Dict [str , float ]] = Field (None )
189
- user : Optional [str ] = Field (None )
190
171
191
172
# llama.cpp specific parameters
192
173
repeat_penalty : float = 1.1
@@ -220,12 +201,7 @@ def create_chat_completion(
220
201
completion_or_chunks = llama .create_chat_completion (
221
202
** request .dict (
222
203
exclude = {
223
- "model" ,
224
- "n" ,
225
- "presence_penalty" ,
226
- "frequency_penalty" ,
227
- "logit_bias" ,
228
- "user" ,
204
+ "model"
229
205
}
230
206
),
231
207
)
0 commit comments