Closed
Description
Name and Version
$ llama-server --version
version: 5627 (3678b83)
built with cc (Ubuntu 14.2.0-19ubuntu2) 14.2.0 for x86_64-linux-gnu
Operating systems
Linux
Which llama.cpp modules do you know to be affected?
llama-server
Command line
llama-server --mmproj /home/{username}/Downloads/models/llama.cpp/bartowski/google_gemma-3-4b-it-GGUF/mmproj-google_gemma-3-4b-it-bf16.gguf -m /home/{username}/Downloads/models/llama.cpp/bartowski/google_gemma-3-4b-it-GGUF/google_gemma-3-4b-it-Q8_0.gguf -t 28 --port 9090 --host 0.0.0.0 -c 0
Problem description & steps to reproduce
To reproduce send 11 images to the server. It's giving me a 500 error on the 11th.
10 images seem to work as expected.
It seemed to allow for many more images (20+) when llama-server was first made multimodal. I'm not sure when this changed.
First Bad Commit
No response
Relevant log output
failed to decode image
failed to decode image decode: failed to find a memory slot for batch of size 256
srv process_chun: image processed in 31744 ms
mtmd_helper_eval failed with status 1slot update_slots: id 0 | task 840 | failed to process image, res = 1
slot release: id 0 | task 840 | stop processing: n_past = 2634, truncated = 0
srv send_error: task id = 840, error: failed to process image