Model outputs only "!!!!!!!!"

#3
by groxaxo - opened

I fire up vLLM, the model loads beautifully across my dual RTX 3090s. Everything looks perfect. The weights are loaded, the KV cache is allocated, the stars are aligned. I send my first prompt: "Hello! How are you today?" The model responds: !!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!! Okay, maybe a bad prompt. Let me try something more elaborate. "Write a Python function to calculate the Fibonacci sequence." !!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!! "Explain quantum physics to a 5-year-old." !!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!! I tried every prompt I could think of β€” creative writing, coding, math, philosophy. The model had one answer for everything, and that answer was exclamation marks. Not even good exclamation marks. Just the same character, repeated infinitely, like a stuck typewriter having an existential crisis. ## Technical details - Model: cyankiwi/Qwen3.5-27B-AWQ-BF16-INT8 - Backend: vLLM 0.19.1rc1 (also tested on dev292) - Hardware: 2x RTX 3090 (48GB total VRAM) - Symptom: Model loads successfully, but every single prompt produces only !!! characters - Quantization: Attempted with --quantization compressed-tensors --enforce-eager β€” same result -

groxaxo changed discussion title from Model outputs only "!!!!!!!!" β€” grandma's hotspot data died for this## to Model outputs only "!!!!!!!!"

Sign up or log in to comment