Michael Yang
de4496b40d
embed templates
2023-07-06 11:33:38 -07:00
Bruce MacDonald
d41e9ee017
display pull progress
2023-07-06 14:18:40 -04:00
Michael Yang
fed04d3255
free llama model
2023-07-06 11:14:32 -07:00
Michael Yang
4b269a7b34
use prompt templates
2023-07-06 10:45:11 -07:00
Bruce MacDonald
92d520b077
pull models
2023-07-06 12:25:00 -04:00
Michael Yang
d24b2a8c58
enable metal gpu acceleration
...
ggml-metal.metal must be in the same directory as the ollama binary
otherwise llama.cpp will not be able to find it and load it.
1. go generate llama/llama_metal.go
2. go build .
3. ./ollama serve
2023-07-06 10:37:05 -04:00
Jeffrey Morgan
f61b221cf9
remove debug print statements
2023-07-06 10:37:05 -04:00
Bruce MacDonald
20598c18d7
separate routes
2023-07-06 10:37:05 -04:00
Jeffrey Morgan
cd3b7e9bac
client updates
2023-07-06 10:36:58 -04:00
Jeffrey Morgan
f1e5a55101
move prompt templates out of python bindings
2023-07-06 10:31:18 -04:00
Jeffrey Morgan
9448e62f38
add llama.cpp go bindings
2023-07-06 10:31:18 -04:00
Jeffrey Morgan
371a9cd9b2
wip go engine
...
Co-authored-by: Patrick Devine <pdevine@sonic.net>
2023-07-06 10:31:18 -04:00