mirror of
https://github.com/tcsenpai/agenticSeek.git
synced 2025-06-03 01:30:11 +00:00
40 lines
1.4 KiB
Python
40 lines
1.4 KiB
Python
|
|
from sources.tools import PyInterpreter, BashInterpreter
|
|
from sources.utility import pretty_print
|
|
from sources.agent import Agent
|
|
|
|
class CoderAgent(Agent):
|
|
def __init__(self, model, name, prompt_path, provider):
|
|
super().__init__(model, name, prompt_path, provider)
|
|
self.set_system_prompt(prompt_path)
|
|
self._tools = {
|
|
"bash": BashInterpreter(),
|
|
"python": PyInterpreter()
|
|
}
|
|
|
|
def answer(self, prompt, speech_module) -> str:
|
|
answer = ""
|
|
attempt = 0
|
|
max_attempts = 3
|
|
self.add_to_history('user', prompt)
|
|
|
|
while attempt < max_attempts:
|
|
pretty_print("Thinking...", color="status")
|
|
self.wait_message(speech_module)
|
|
answer, reasoning = self.llm_request(self.history)
|
|
exec_success, feedback = self.execute_modules(answer)
|
|
pretty_print(feedback, color="failure" if "failure" in feedback.lower() else "success")
|
|
if exec_success:
|
|
break
|
|
attempt += 1
|
|
|
|
return answer, reasoning
|
|
|
|
if __name__ == "__main__":
|
|
from llm_provider import Provider
|
|
|
|
#local_provider = Provider("ollama", "deepseek-r1:14b", None)
|
|
server_provider = Provider("server", "deepseek-r1:14b", "192.168.1.100:5000")
|
|
agent = CoderAgent("deepseek-r1:14b", "jarvis", "prompts/coder_agent.txt", server_provider)
|
|
ans = agent.answer("What is the output of 5+5 in python ?")
|
|
print(ans) |