|
ext_server
|
llm: add back check for empty token cache
|
2024-04-30 17:38:44 -04:00 |
|
generate
|
Do not build AVX runners on ARM64
|
2024-04-26 23:55:32 -06:00 |
|
llama.cpp@952d03dbea
|
update llama.cpp commit to `952d03d`
|
2024-04-30 17:31:20 -04:00 |
|
patches
|
Fix clip log import
|
2024-04-26 09:43:46 -07:00 |
|
filetype.go
|
comments
|
2024-05-02 17:05:26 -07:00 |
|
ggla.go
|
split model layer into metadata and data layers
|
2024-05-02 17:05:49 -07:00 |
|
ggml.go
|
split model layer into metadata and data layers
|
2024-05-02 17:05:49 -07:00 |
|
gguf.go
|
split model layer into metadata and data layers
|
2024-05-02 17:05:49 -07:00 |
|
llm.go
|
comments
|
2024-05-02 17:05:26 -07:00 |
|
llm_linux.go
|
Switch back to subprocessing for llama.cpp
|
2024-04-01 16:48:18 -07:00 |
|
server.go
|
Removing go routine calling .wait from load.
|
2024-05-01 18:51:10 +00:00 |
|
status.go
|
Switch back to subprocessing for llama.cpp
|
2024-04-01 16:48:18 -07:00 |