ollama/model
Jesse Gross 4d4463b2bd backend: Support graph computation that does not return an output
There are two cases where we may not have an output after computing:
 - Prompt processing where the length of the input exceeds the batch
   size
 - Internal memory management operations such as cache defrag and shift
2025-02-13 17:09:26 -08:00
..
imageproc imageproc mllama refactor (#7537) 2024-12-14 19:50:15 -08:00
llama backend: Consistently use int (vs. int64) for tensor shapes 2025-02-13 17:09:26 -08:00
mllama backend: Consistently use int (vs. int64) for tensor shapes 2025-02-13 17:09:26 -08:00
pixtral imageproc mllama refactor (#7537) 2024-12-14 19:50:15 -08:00
qwen2vl imageproc mllama refactor (#7537) 2024-12-14 19:50:15 -08:00
testdata next ollama runner (#7913) 2025-02-13 16:31:21 -08:00
model.go backend: Support graph computation that does not return an output 2025-02-13 17:09:26 -08:00
model_test.go next ollama runner (#7913) 2025-02-13 16:31:21 -08:00
process_text.go next ollama runner (#7913) 2025-02-13 16:31:21 -08:00
process_text_test.go next ollama runner (#7913) 2025-02-13 16:31:21 -08:00