ollama/model
Michael Yang 9213339549
fix: qwen25vl assign samebatch in multimodal input (#10789)
setting samebatch on the vision start token is problematic because it
will be shared with other inputs that also use images. this will cause
the input to be cached and the runner will not see SameBatch. SameBatch
will also be incorrect since it may be for a different image.

assigning samebatch to the input tokens resolves this by ensure it's
assigned correctly to inputs corresponding to the image.

not setting same batch correctly may cause panics during inference since
images are no longer guaranteed to be in the same batch.
2025-12-29 06:38:03 -06:00
..
imageproc imageproc mllama refactor (#7537) 2024-12-14 19:50:15 -08:00
input ollamarunner: Separate text and multimodal graphs 2025-12-29 06:38:01 -06:00
models fix: qwen25vl assign samebatch in multimodal input (#10789) 2025-12-29 06:38:03 -06:00
testdata gemma2 impl 2025-03-11 14:35:08 -07:00
bytepairencoding.go model: handle multiple eos tokens (#10577) 2025-12-29 06:38:01 -06:00
bytepairencoding_test.go model: handle multiple eos tokens (#10577) 2025-12-29 06:38:01 -06:00
model.go ggml: Seperate tensor load from backend creation 2025-12-29 06:38:02 -06:00
model_test.go fs: move ml.Config to fs package 2025-04-03 13:12:24 -07:00
sentencepiece.go model: handle multiple eos tokens (#10577) 2025-12-29 06:38:01 -06:00
sentencepiece_test.go model: handle multiple eos tokens (#10577) 2025-12-29 06:38:01 -06:00
textprocessor.go model: handle multiple eos tokens (#10577) 2025-12-29 06:38:01 -06:00
vocabulary.go model: handle multiple eos tokens (#10577) 2025-12-29 06:38:01 -06:00