ollama/runner/ollamarunner
Jesse Gross 06007c0a18 Allow models to force a new batch
This is useful for a few things:
 - Work around bugs, such as having 2 images in one batch
 - Keep the image in a single batch for fully connected attention
 - Improve performance by not evaluating embeddings multiple times
2025-03-11 14:49:20 -07:00
..
cache.go model: Update encoder cache to use multimodal input processing handler 2025-03-09 17:05:26 -07:00
cache_test.go model: Update encoder cache to use multimodal input processing handler 2025-03-09 17:05:26 -07:00
runner.go Allow models to force a new batch 2025-03-11 14:49:20 -07:00