ollama/llm
Daniel Hiltgen 39ca55a1ba
Move quantization to new backend (#10363)
* Move quantization logic to GGML via new backend

This moves the model aware logic to Go code and calls GGMLs quantization code for model creation.

* Remove "add model quantizations"

This is no longer needed now that quantization is implemented in Go+GGML code directly.
2025-12-29 06:37:52 -06:00
..
llm_darwin.go Optimize container images for startup (#6547) 2024-09-12 12:10:30 -07:00
llm_linux.go Optimize container images for startup (#6547) 2024-09-12 12:10:30 -07:00
llm_windows.go win: lint fix (#10571) 2025-12-29 06:37:51 -06:00
memory.go explicitly decode maxarraysize 1024 2025-12-29 06:37:44 -06:00
memory_test.go Move quantization to new backend (#10363) 2025-12-29 06:37:52 -06:00
server.go api: remove unused or unsupported api options (#10574) 2025-12-29 06:37:52 -06:00
server_test.go llm: do not error on "null" format (#8139) 2024-12-17 09:49:37 -08:00
status.go Improve crash reporting (#7728) 2024-11-19 16:26:57 -08:00