Very buggy in vllm + opencode

#40
by void009 - opened

I am hitting at least this bug in vllm: https://github.com/vllm-project/vllm/issues/39056
using the vllm settings in you model card (with TP).
It often stops while doing tool call in thinking mode, sometimes i see just


and it stops.

Sometimes I get "The question tool was called with invalid arguments, expected 'string', received 'undefined'.
Other times bad json errors.

This is with full fp16.
I tried tool parsing qwen3_coder as well as qwen3_xml, as well as the patch mentioned in the github issue, none helped. I've been attempting to make it run for a few days now. I also attempted the preserve thinking template override. Using openai compatible API.

Can someone from qwen take a look and see if there's any possible working config with vllm + opencode ?

yes having the same issues.

Sign up or log in to comment