Commit Graph

21 Commits

Author SHA1 Message Date
Michael Yang de4496b40d embed templates 2023-07-06 11:33:38 -07:00
Bruce MacDonald d41e9ee017 display pull progress 2023-07-06 14:18:40 -04:00
Michael Yang fed04d3255 free llama model 2023-07-06 11:14:32 -07:00
Michael Yang 4b269a7b34 use prompt templates 2023-07-06 10:45:11 -07:00
Bruce MacDonald 92d520b077 pull models 2023-07-06 12:25:00 -04:00
Michael Yang d24b2a8c58 enable metal gpu acceleration
ggml-metal.metal must be in the same directory as the ollama binary
otherwise llama.cpp will not be able to find it and load it.

1. go generate llama/llama_metal.go
2. go build .
3. ./ollama serve
2023-07-06 10:37:05 -04:00
Jeffrey Morgan f61b221cf9 remove debug print statements 2023-07-06 10:37:05 -04:00
Bruce MacDonald 20598c18d7 separate routes 2023-07-06 10:37:05 -04:00
Jeffrey Morgan cd3b7e9bac client updates 2023-07-06 10:36:58 -04:00
Jeffrey Morgan f1e5a55101 move prompt templates out of python bindings 2023-07-06 10:31:18 -04:00
Jeffrey Morgan 9448e62f38 add llama.cpp go bindings 2023-07-06 10:31:18 -04:00
Jeffrey Morgan 371a9cd9b2 wip go engine
Co-authored-by: Patrick Devine <pdevine@sonic.net>
2023-07-06 10:31:18 -04:00
Jeffrey Morgan b361fa72ec reorganize directories 2023-06-25 13:08:03 -04:00
Jeffrey Morgan d3709f85b5 build server into desktop app 2023-06-25 00:30:02 -04:00
Jeffrey Morgan 369108e1ad Add dependencies to `requirements.txt` 2023-06-24 21:13:26 -04:00
Michael Chiang 5142ba2dad
Update README.md 2023-06-23 22:51:54 -04:00
Bruce MacDonald c5bafaff54 package server with client 2023-06-23 18:38:22 -04:00
Bruce MacDonald f0eee3faa0 build server executable 2023-06-23 17:23:30 -04:00
Bruce MacDonald ebec1c61db load and unload model endpoints 2023-06-23 14:47:57 -04:00
Bruce MacDonald 0758cb2d4b llama server wrapper 2023-06-23 13:10:13 -04:00
Jeffrey Morgan 8fa91332fa initial commit 2023-06-22 18:31:40 -04:00