Get up and running with OpenAI gpt-oss, DeepSeek-R1, Gemma 3 and other models.
Go to file
Jeffrey Morgan 0389affa07 wip 2023-07-04 16:06:04 -04:00
api wip 2023-07-04 16:06:04 -04:00
app move .prettierrc.json to root 2023-07-02 17:34:46 -04:00
cmd wip 2023-07-04 16:06:04 -04:00
docs Move python docs to separate file 2023-07-01 17:54:29 -04:00
llama wip 2023-07-04 16:06:04 -04:00
server wip 2023-07-04 00:47:00 -04:00
templates move prompt templates out of python bindings 2023-07-03 17:14:20 -04:00
web use app router 2023-07-02 22:05:26 -04:00
.dockerignore update `Dockerfile` 2023-07-03 18:56:17 -04:00
.gitignore add templates to prompt command 2023-06-26 13:41:16 -04:00
.prettierrc.json move .prettierrc.json to root 2023-07-02 17:34:46 -04:00
Dockerfile update `Dockerfile` 2023-07-03 18:56:17 -04:00
LICENSE `proto` -> `ollama` 2023-06-26 15:57:13 -04:00
README.md add llama.cpp go bindings 2023-07-03 16:32:48 -04:00
go.mod wip 2023-07-04 16:06:04 -04:00
go.sum wip 2023-07-04 16:06:04 -04:00
main.go add llama.cpp go bindings 2023-07-03 16:32:48 -04:00
models.json format `models.json` 2023-07-02 20:33:23 -04:00

README.md

Ollama

An easy, fast runtime for large language models, powered by llama.cpp.

Note: this project is a work in progress. Certain models that can be run with ollama are intended for research and/or non-commercial use only.

Install

Using pip:

pip install ollama

Using docker:

docker run ollama/ollama

Quickstart

To run a model, use ollama run:

ollama run orca-mini-3b

You can also run models from hugging face:

ollama run huggingface.co/TheBloke/orca_mini_3B-GGML

Or directly via downloaded model files:

ollama run ~/Downloads/orca-mini-13b.ggmlv3.q4_0.bin

Building

go generate ./...
go build .

Documentation