agenticSeek/README.md
2025-03-06 11:32:40 +01:00

110 lines
2.5 KiB
Markdown
Raw Blame History

This file contains invisible Unicode characters

This file contains invisible Unicode characters that are indistinguishable to humans but may be processed differently by a computer. If you think that this is intentional, you can safely ignore this warning. Use the Escape button to reveal them.

This file contains Unicode characters that might be confused with other characters. If you think that this is intentional, you can safely ignore this warning. Use the Escape button to reveal them.

# 🚀 agenticSeek: Local AI Assistant Powered by DeepSeek Agents
**A fully local AI assistant** using Deepseek R1 agents.
> 🛠️ **Work in Progress** Looking for contributors! 🚀
---
## Features:
- **Privacy-first**: Runs 100% locally **no data leaves your machine**
- **Voice-enabled**: Speak and interact naturally
- **Coding abilities**: Code in Python, Bash, C, Golang, and soon more
- **Self-correcting**: Automatically fixes errors by itself
- **Agent routing**: Select the best agent for the task
- **Multi-agent**: For complex tasks, divide and conquer with multiple agents
- **Web browsing (not implemented yet)**: Browse the web and search the internet
---
## Installation
### 1⃣ **Install Dependencies**
```sh
pip3 install -r requirements.txt
```
### 2⃣ **Download Models**
Make sure you have [Ollama](https://ollama.com/) installed.
Download the `deepseek-r1:7b` model from [DeepSeek](https://deepseek.com/models)
```sh
ollama pull deepseek-r1:7b
```
### 3⃣ **Run the Assistant (Ollama)**
Start the ollama server
```sh
ollama serve
```
Change the config.ini file to set the provider_name to `ollama` and provider_model to `deepseek-r1:7b`
```sh
[MAIN]
is_local = True
provider_name = ollama
provider_model = deepseek-r1:7b
```
Run the assistant:
```sh
python3 main.py
```
### 4⃣ **Alternative: Run the Assistant (Own Server)**
Get the ip address of the machine that will run the model
```sh
ip a | grep "inet " | grep -v 127.0.0.1 | awk '{print $2}' | cut -d/ -f1
```
On the other machine that will run the model execute the script in stream_llm.py
```sh
python3 stream_llm.py
```
Change the `config.ini` file to set the `provider_name` to `server` and `provider_model` to `deepseek-r1:7b`.
Set the `provider_server_address` to the ip address of the machine that will run the model.
```sh
[MAIN]
is_local = False
provider_name = server
provider_model = deepseek-r1:14b
provider_server_address = x.x.x.x:5000
```
Run the assistant:
```sh
python3 main.py
```
## Current capabilities
- All running locally
- Reasoning with deepseek R1
- Code execution capabilities (Python, Golang, C)
- Shell control capabilities in bash
- Will try to fix errors by itself
- Routing system, select the best agent for the task
- Fast text-to-speech using kokoro.
- Memory compression (reduce history as interaction progresses using summary model)
- Recovery: recover last session from memory
## UNDER DEVELOPMENT
- Web browsing
- Knowledge base RAG
- Graphical interface
- Speech-to-text using distil-whisper/distil-medium.en