From 8ed5fa7a35079f4354472ab95665af439d007d95 Mon Sep 17 00:00:00 2001 From: Martin <49105846+Fosowl@users.noreply.github.com> Date: Mon, 10 Mar 2025 14:41:07 +0100 Subject: [PATCH] Update README.md --- README.md | 24 +++++++++++++++++++++++- 1 file changed, 23 insertions(+), 1 deletion(-) diff --git a/README.md b/README.md index 7850f88..75c0be1 100644 --- a/README.md +++ b/README.md @@ -24,7 +24,7 @@ --- -## Run locally +## Run locally on your machine **We recommend using at least Deepseek 14B, smaller models struggle with tool use and forget quickly the context.** @@ -67,6 +67,8 @@ python3 main.py ## **Alternative: Run the LLM on your own server** +If you have a powerful computer or a server that you can use, but you want to use it from your laptop you have the options to run the LLM on a remote server. + ### 1️⃣ **Set up and start the server scripts** On your "server" that will run the AI model, get the ip address @@ -104,6 +106,26 @@ Run the assistant: python3 main.py ``` +## **Run with an API** + +Clone the repository. + +Set the desired provider in the `config.ini` + +```sh +[MAIN] +is_local = False +provider_name = openai +provider_model = gpt4-o +provider_server_address = 127.0.0.1:5000 # can be set to anything, not used +``` + +Run the assistant: + +```sh +python3 main.py +``` + ## Providers The table below show the available providers: