ollama/model/models/gemma3
Jesse Gross 282bfaaa95 ollamarunner: Use a separate context per multimodal input
Currently there is a single context per sequence, shared all by
all multimodal inputs. Since we build a vision encoder graph per
image, with a large number of inputs we can eventually hit the
maximum number of graph nodes per context.

This changes to use a separate context for each image, ensuring
that available resource limits are consistent.
2025-03-14 15:38:54 -07:00
..
model.go ollamarunner: Use a separate context per multimodal input 2025-03-14 15:38:54 -07:00
model_text.go ml: Allow models to constrain inputs to a single batch 2025-03-14 15:38:54 -07:00
model_vision.go all: address linter errors 2025-03-11 14:49:20 -07:00
process_image.go fix vision encoder 2025-03-11 14:49:19 -07:00