Skip to content

Commit 0a91cc3

Browse files
committed
minor fix
1 parent cd9f058 commit 0a91cc3

File tree

2 files changed

+7
-5
lines changed

2 files changed

+7
-5
lines changed

llama_cpp/server/app.py

Lines changed: 6 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -49,9 +49,10 @@
4949

5050
_server_settings: Optional[ServerSettings] = None
5151

52-
DEFAULT_GRAMMAR_FILE = os.getenv("DEFAULT_GRAMMAR_FILE", "./json_grammar.gbnf")
53-
with open(DEFAULT_GRAMMAR_FILE) as f:
54-
default_grammar_text = f.read()
52+
DEFAULT_GRAMMAR_FILE = os.getenv("GRAMMAR_FILE", "")
53+
if DEFAULT_GRAMMAR_FILE != "":
54+
with open(DEFAULT_GRAMMAR_FILE) as f:
55+
default_grammar_text = f.read()
5556

5657
def set_server_settings(server_settings: ServerSettings):
5758
global _server_settings
@@ -308,7 +309,8 @@ async def create_chat_completion(
308309
if body.grammar is not None:
309310
kwargs["grammar"] = llama_cpp.LlamaGrammar.from_string(body.grammar)
310311
else:
311-
kwargs["grammar"] = llama_cpp.LlamaGrammar.from_string(default_grammar_text)
312+
if DEFAULT_GRAMMAR_FILE:
313+
kwargs["grammar"] = llama_cpp.LlamaGrammar.from_string(default_grammar_text)
312314
iterator_or_completion: Union[
313315
llama_cpp.ChatCompletion, Iterator[llama_cpp.ChatCompletionChunk]
314316
] = await run_in_threadpool(llama.create_chat_completion, **kwargs)

vendor/llama.cpp

0 commit comments

Comments
 (0)