From a667f89c124569983f25e00df70708495c0fd33e Mon Sep 17 00:00:00 2001 From: martin legrand Date: Sat, 5 Apr 2025 16:41:24 +0200 Subject: [PATCH] fix : server cache --- server/sources/ollama_handler.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/server/sources/ollama_handler.py b/server/sources/ollama_handler.py index 8f2d43c..517c284 100644 --- a/server/sources/ollama_handler.py +++ b/server/sources/ollama_handler.py @@ -13,8 +13,8 @@ class OllamaLLM(GeneratorLLM): def generate(self, history): self.logger.info(f"Using {self.model} for generation with Ollama") - if cache.is_cached(history[-1]['content']): - self.state.current_buffer = cache.get_cached_response(history[-1]['content']) + if self.cache.is_cached(history[-1]['content']): + self.state.current_buffer = self.cache.get_cached_response(history[-1]['content']) self.state.is_generating = False return try: