llama.cpp flags / visual token budget
#26
by 234r89r23u89023rui90 - opened
Hi,
thanks for keeping the GGUFs up to date. How to properly set the visual token budget [0] in llama.cpp server cli or api requests?
Appart from that, is using --jinja not needed anymore? (asking because you do not use it in your examples)
[0] https://unsloth.ai/docs/models/gemma-4#ocr-document-prompt