update readme

This commit is contained in:
Roy Han 2024-08-09 11:32:27 -07:00
parent 89f3bae306
commit 30823ec925

View File

@ -2,15 +2,24 @@
### To run
`make {/path/to/whisper.cpp/server}`
- replace `whisperServer` in `routes.go` with path to server
## CLI
`./ollama run llama3 [PROMPT] --speech`
- processes voice audio with the provided prompt
`./ollama run llama3 --speech`
- enters interactive mode for continuous voice chat
- TODO: fix exiting interactive mode
Notes: uses default model
### Update routes.go
- replace `whisperServer` with path to server
## api/generate
### Request fields
- `speech` (required):
- `audio` (required): path to audio file
- `model` (required): path to whisper model
- `model` (optional): path to whisper model, uses default if null
- `transcribe` (optional): if true, will transcribe and return the audio file
- `keep_alive`: (optional): sets how long the model is stored in memory (default: `5m`)
- `prompt` (optional): if not null, passed in with the transcribed audio
@ -45,9 +54,10 @@ curl http://localhost:11434/api/generate -d '{
## api/chat
### Request fields
- `model` (required): language model to chat with
- `speech` (required):
- `model` (required): path to whisper model
- `speech` (optional):
- `model` (optional): path to whisper model, uses default if null
- `keep_alive`: (optional): sets how long the model is stored in memory (default: `5m`)
- `run_speech` (optional): either this flag must be true or `speech` must be passed in for speech mode to run
- `messages`/`message`/`audio` (required): path to audio file
```