Very buggy in vllm + opencode
I am hitting at least this bug in vllm: https://github.com/vllm-project/vllm/issues/39056
using the vllm settings in you model card (with TP).
It often stops while doing tool call in thinking mode, sometimes i see just
and it stops.
Sometimes I get "The question tool was called with invalid arguments, expected 'string', received 'undefined'.
Other times bad json errors.
This is with full fp16.
I tried tool parsing qwen3_coder as well as qwen3_xml, as well as the patch mentioned in the github issue, none helped. I've been attempting to make it run for a few days now. I also attempted the preserve thinking template override. Using openai compatible API.
Can someone from qwen take a look and see if there's any possible working config with vllm + opencode ?
yes having the same issues.