- remove new lines from llama.cpp error messages relayed to client - check api option types and return error on wrong type - change num layers from 95% VRAM to 92% VRAM |
||
|---|---|---|
| .. | ||
| llama.cpp | ||
| falcon.go | ||
| ggml.go | ||
| gguf.go | ||
| llama.go | ||
| llm.go | ||
| starcoder.go | ||
| utils.go | ||