3297 Commits

Author SHA1 Message Date
jmorganca
57d03929cd fix build on windows 2024-07-29 15:38:51 -07:00
jmorganca
0a6b1adbd7 fix ggml-metal.m build constraints 2024-07-29 15:38:51 -07:00
jmorganca
ec60d79a67 fix ggml-metal.m 2024-07-29 15:38:51 -07:00
jmorganca
3d656588a7 avx2 should only add avx2 2024-07-29 15:38:51 -07:00
jmorganca
460d9857e2 fix sync script 2024-07-29 15:38:51 -07:00
jmorganca
a5548a81fc fix ggml-metal.m 2024-07-29 15:38:51 -07:00
jmorganca
634f6a75d0 fix ggml-metal.m 2024-07-29 15:38:51 -07:00
jmorganca
3b5e5a6280 add license headers 2024-07-29 15:38:51 -07:00
jmorganca
853d96b1b1 pre-patch 2024-07-29 15:38:51 -07:00
jmorganca
4dd63c1fef move runner package down 2024-07-29 15:38:51 -07:00
jmorganca
82214396b5 replace static build in llm 2024-07-29 15:38:51 -07:00
jmorganca
8ca4a9a70a fix build 2024-07-29 15:35:09 -07:00
jmorganca
25fd8fd045 wip... 2024-07-29 15:35:09 -07:00
jmorganca
be2f37b5d4 rename server to runner 2024-07-29 15:35:09 -07:00
Jeffrey Morgan
9e28405c54 Update README.md 2024-07-29 15:35:09 -07:00
Jeffrey Morgan
9f3e950120 Update README.md 2024-07-29 15:35:09 -07:00
Jeffrey Morgan
951104045f Update README.md 2024-07-29 15:35:09 -07:00
Jeffrey Morgan
597712006c Update README.md 2024-07-29 15:35:09 -07:00
jmorganca
64e712b12b Add missing hipcc flags 2024-07-29 15:35:09 -07:00
jmorganca
85aea62997 fix .gitattributes 2024-07-29 15:35:09 -07:00
jmorganca
491ff41675 Initial llama Go module 2024-07-29 15:35:09 -07:00
jmorganca
075f2e88d9 add sync of llama.cpp 2024-07-29 15:35:09 -07:00
Daniel Hiltgen
1a83581a8e
Merge pull request #5895 from dhiltgen/sched_faq
Better explain multi-gpu behavior
2024-07-29 14:25:41 -07:00
Daniel Hiltgen
37926eb991
Merge pull request #5927 from dhiltgen/high_cpu_count
Ensure amd gpu nodes are numerically sorted
2024-07-29 14:24:57 -07:00
Daniel Hiltgen
3d4634fdff
Merge pull request #5934 from dhiltgen/missing_cuda_repo
Report better error on cuda unsupported os/arch
2024-07-29 14:24:20 -07:00
royjhan
365431d406
return tool calls finish reason for openai (#5995)
* hot fix

* backend stream support

* clean up

* finish reason

* move to openai
2024-07-29 13:56:57 -07:00
Daniel Hiltgen
161e12cecf
Merge pull request #5932 from dhiltgen/win_font
Explain font problems on windows 10
2024-07-29 13:40:24 -07:00
Jeffrey Morgan
46e6327e0f
api: add stringifier for Tool (#5891) 2024-07-29 13:35:16 -07:00
Jeffrey Morgan
68ee42f995
update llama.cpp submodule to 6eeaeba1 (#6039) 2024-07-29 13:20:26 -07:00
Ikko Eltociear Ashimine
f26aef9a8b
docs: update README.md (#6059)
HuggingFace -> Hugging Face
2024-07-29 10:53:30 -07:00
Michael Yang
38d9036b59
Merge pull request #5992 from ollama/mxyng/save
fix: model save
2024-07-29 09:53:19 -07:00
Veit Heller
6f26e9322f
Fix typo in image docs (#6041) 2024-07-29 08:50:53 -07:00
Jeffrey Morgan
0e4d653687
upate to llama3.1 elsewhere in repo (#6032) 2024-07-28 19:56:02 -07:00
Michael
2c01610616
update readme to llama3.1 (#5933) 2024-07-28 14:21:38 -07:00
Tibor Schmidt
f3d7a481b7
feat: add support for min_p (resolve #1142) (#1825) 2024-07-27 14:37:40 -07:00
Jeffrey Morgan
f2a96c7d77
llm: keep patch for llama 3 rope factors (#5987) 2024-07-26 15:20:52 -07:00
Daniel Hiltgen
e8a66680d1
Merge pull request #5705 from dhiltgen/win_errormode
Enable windows error dialog for subprocess
2024-07-26 14:49:34 -07:00
Michael Yang
079b2c3b03
Merge pull request #5999 from ollama/mxyng/fix-push
fix nil deref in auth.go
2024-07-26 14:28:34 -07:00
Blake Mizerany
750c1c55f7
server: fix race conditions during download (#5994)
This fixes various data races scattered throughout the download/pull
client where the client was accessing the download state concurrently.

This commit is mostly a hot-fix and will be replaced by a new client one
day soon.

Also, remove the unnecessary opts argument from downloadChunk.
2024-07-26 14:24:24 -07:00
Michael Yang
a622c47bd3 fix nil deref in auth.go 2024-07-26 14:14:48 -07:00
Michael Yang
ec4c35fe99
Merge pull request #5512 from ollama/mxyng/detect-stop
autodetect stop parameters from template
2024-07-26 13:48:23 -07:00
Michael Yang
3d9de805b7 fix: model save
stop parameter is saved as a slice which is incompatible with modelfile
parsing
2024-07-26 13:23:06 -07:00
Jeffrey Morgan
f5e3939220
Update api.md (#5968) 2024-07-25 23:10:18 -04:00
Jeffrey Morgan
ae27d9dcfd
Update openai.md 2024-07-25 20:27:33 -04:00
Michael Yang
37096790a7
Merge pull request #5552 from ollama/mxyng/messages-docs
docs
2024-07-25 16:26:19 -07:00
Michael Yang
997c903884
Update docs/template.md
Co-authored-by: Jeffrey Morgan <jmorganca@gmail.com>
2024-07-25 16:23:40 -07:00
Blake Mizerany
c8af3c2d96
server: reuse original download URL for images (#5962)
This changes the registry client to reuse the original download URL
it gets on the first redirect response for all subsequent requests,
preventing thundering herd issues when hot new LLMs are released.
2024-07-25 15:58:30 -07:00
Jeffrey Morgan
455e61170d
Update openai.md 2024-07-25 18:34:47 -04:00
royjhan
4de1370a9d
openai tools doc (#5617) 2024-07-25 18:34:06 -04:00
Jeffrey Morgan
bbf8f102ee
Revert "llm(llama): pass rope factors (#5924)" (#5963)
This reverts commit bb46bbcf5e90e5efab5ff946a6c798131907ba2d.
v0.3.0
2024-07-25 18:24:55 -04:00