From ddb533a255951840970c2b4542de3eeeede7218b Mon Sep 17 00:00:00 2001 From: martin legrand Date: Sat, 29 Mar 2025 19:27:08 +0100 Subject: [PATCH] fix : server model download --- server/sources/llamacpp_handler.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/server/sources/llamacpp_handler.py b/server/sources/llamacpp_handler.py index 758be3f..5bc108e 100644 --- a/server/sources/llamacpp_handler.py +++ b/server/sources/llamacpp_handler.py @@ -13,12 +13,12 @@ class LlamacppLLM(GeneratorLLM): def generate(self, history): if self.llm is None: + self.logger.info(f"Loading {self.model}...") self.llm = Llama.from_pretrained( repo_id=self.model, filename="*Q8_0.gguf", verbose=True ) - return self.logger.info(f"Using {self.model} for generation with Llama.cpp") try: with self.state.lock: