mirror of
https://github.com/tcsenpai/agenticSeek.git
synced 2025-06-06 19:15:28 +00:00
Update README.md
This commit is contained in:
parent
1433aa50aa
commit
b114cc8cdc
41
README.md
41
README.md
@ -4,17 +4,15 @@
|
|||||||
|
|
||||||
**A fully local alternative to Manus AI**, a voice-enabled AI assistant that codes, explores your filesystem, browse the web and correct it's mistakes all without sending a byte of data to the cloud. Built with reasoning models like DeepSeek R1, this autonomous agent runs entirely on your hardware, keeping your data private.
|
**A fully local alternative to Manus AI**, a voice-enabled AI assistant that codes, explores your filesystem, browse the web and correct it's mistakes all without sending a byte of data to the cloud. Built with reasoning models like DeepSeek R1, this autonomous agent runs entirely on your hardware, keeping your data private.
|
||||||
|
|
||||||
|
[](https://fosowl.github.io/agenticSeek.html)  
|
||||||
> 🛠️ **Work in Progress** – Looking for contributors!
|
> 🛠️ **Work in Progress** – Looking for contributors!
|
||||||
|
|
||||||

|

|
||||||
|
|
||||||
|
|
||||||
[](https://fosowl.github.io/agenticSeek.html)   
|
|
||||||
|
|
||||||
|
|
||||||
---
|
|
||||||
|
|
||||||
## Features:
|
## Features:
|
||||||
|
|
||||||
- **100% Local**: No cloud, runs on your hardware. Your data stays yours.
|
- **100% Local**: No cloud, runs on your hardware. Your data stays yours.
|
||||||
|
|
||||||
- **Voice interaction**: Voice-enabled natural interaction.
|
- **Voice interaction**: Voice-enabled natural interaction.
|
||||||
@ -37,16 +35,33 @@
|
|||||||
|
|
||||||
---
|
---
|
||||||
|
|
||||||
## Run locally on your machine
|
## **Installation**
|
||||||
|
|
||||||
**We recommend using at least Deepseek 14B, smaller models struggle with tool use and forget quickly the context.**
|
### 1️⃣ **Clone the repository**
|
||||||
|
|
||||||
|
```sh
|
||||||
|
git clone https://github.com/Fosowl/agenticSeek.git
|
||||||
|
cd agenticSeek
|
||||||
|
```
|
||||||
|
|
||||||
|
### 2️ **Create a virtual env**
|
||||||
|
|
||||||
|
```sh
|
||||||
|
python3 -m venv agentic_seek_env
|
||||||
|
source agentic_seek_env/bin/activate # On Windows: agentic_seek_env\Scripts\activate
|
||||||
|
```
|
||||||
|
|
||||||
|
### 3️⃣ **Install Dependencies**
|
||||||
|
|
||||||
### 1️⃣ **Install Dependencies**
|
|
||||||
```sh
|
```sh
|
||||||
pip3 install -r requirements.txt
|
pip3 install -r requirements.txt
|
||||||
```
|
```
|
||||||
|
|
||||||
### 2️⃣ **Download Models**
|
## Run locally on your machine
|
||||||
|
|
||||||
|
**We recommend using at least Deepseek 14B, smaller models struggle with tool use and forget quickly the context.**
|
||||||
|
|
||||||
|
### 1️⃣ **Download Models**
|
||||||
|
|
||||||
Make sure you have [Ollama](https://ollama.com/) installed.
|
Make sure you have [Ollama](https://ollama.com/) installed.
|
||||||
|
|
||||||
@ -56,7 +71,7 @@ Download the `deepseek-r1:7b` model from [DeepSeek](https://deepseek.com/models)
|
|||||||
ollama pull deepseek-r1:7b
|
ollama pull deepseek-r1:7b
|
||||||
```
|
```
|
||||||
|
|
||||||
### 3️⃣ **Run the Assistant (Ollama)**
|
### 2️ **Run the Assistant (Ollama)**
|
||||||
|
|
||||||
Start the ollama server
|
Start the ollama server
|
||||||
```sh
|
```sh
|
||||||
@ -78,7 +93,9 @@ Run the assistant:
|
|||||||
python3 main.py
|
python3 main.py
|
||||||
```
|
```
|
||||||
|
|
||||||
## **Alternative: Run the LLM on your own server**
|
---
|
||||||
|
|
||||||
|
## **Run the LLM on your own server**
|
||||||
|
|
||||||
If you have a powerful computer or a server that you can use, but you want to use it from your laptop you have the options to run the LLM on a remote server.
|
If you have a powerful computer or a server that you can use, but you want to use it from your laptop you have the options to run the LLM on a remote server.
|
||||||
|
|
||||||
@ -90,6 +107,8 @@ On your "server" that will run the AI model, get the ip address
|
|||||||
ip a | grep "inet " | grep -v 127.0.0.1 | awk '{print $2}' | cut -d/ -f1
|
ip a | grep "inet " | grep -v 127.0.0.1 | awk '{print $2}' | cut -d/ -f1
|
||||||
```
|
```
|
||||||
|
|
||||||
|
Note: For Windows or macOS, use ipconfig or ifconfig respectively to find the IP address.
|
||||||
|
|
||||||
Clone the repository and then, run the script `stream_llm.py` in `server/`
|
Clone the repository and then, run the script `stream_llm.py` in `server/`
|
||||||
|
|
||||||
```sh
|
```sh
|
||||||
@ -139,6 +158,8 @@ Run the assistant:
|
|||||||
python3 main.py
|
python3 main.py
|
||||||
```
|
```
|
||||||
|
|
||||||
|
---
|
||||||
|
|
||||||
## Providers
|
## Providers
|
||||||
|
|
||||||
The table below show the available providers:
|
The table below show the available providers:
|
||||||
|
Loading…
x
Reference in New Issue
Block a user