agenticSeek/server/sources/llamacpp.py
2025-03-29 13:31:13 +01:00

24 lines
643 B
Python

from .generator import GeneratorLLM
class LlamacppLLM(GeneratorLLM):
def __init__(self):
from llama_cpp import Llama
"""
Handle generation using llama.cpp
"""
super().__init__()
self.llm = None
def generate(self, history):
if self.model is None:
self.llm = Llama.from_pretrained(
repo_id=self.model,
filename="*q8_0.gguf",
verbose=True
)
self.logger.info(f"Using {self.model} for generation with Llama.cpp")
self.llm.create_chat_completion(
messages = history
)