llama.cpp flags / visual token budget

#26
by 234r89r23u89023rui90 - opened

Hi,
thanks for keeping the GGUFs up to date. How to properly set the visual token budget [0] in llama.cpp server cli or api requests?
Appart from that, is using --jinja not needed anymore? (asking because you do not use it in your examples)

[0] https://unsloth.ai/docs/models/gemma-4#ocr-document-prompt

Sign up or log in to comment