From 60a17e396f9ab6ad41f92ce709c4240219f7e835 Mon Sep 17 00:00:00 2001 From: Mert Cobanov Date: Fri, 26 Apr 2024 04:34:59 +0300 Subject: [PATCH] Optimize summarizer prompt template and update llama model --- summarizer.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/summarizer.py b/summarizer.py index db12eff..bd59acd 100644 --- a/summarizer.py +++ b/summarizer.py @@ -37,7 +37,7 @@ def setup_summarization_chain(): 5.Conclude your notes with [End of Notes, Message #X] to indicate completion, where "X" represents the total number of messages that I have sent. In other words, include a message counter where you start with #1 and add 1 to the message counter every time I send a message. - By following this optimized prompt, you will generate an effective summary that encapsulates the essence of the given text in a clear, detailed, and reader-friendly manner. + By following this optimized prompt, you will generate an effective summary that encapsulates the essence of the given text in a clear, detailed, and reader-friendly manner. Optimize output as markdown file. "{text}" @@ -45,7 +45,7 @@ def setup_summarization_chain(): input_variables=["text"], ) - llm = ChatOllama(model="llama2") + llm = ChatOllama(model="llama3") llm_chain = LLMChain(llm=llm, prompt=prompt_template) return llm_chain