Hugging Face's logo Hugging Face
  • Models
  • Datasets
  • Spaces
  • Buckets new
  • Docs
  • Enterprise
  • Pricing

  • Log In
  • Sign Up

Qwen
/
Qwen3.5-397B-A17B-FP8

Image-Text-to-Text
Transformers
Safetensors
qwen3_5_moe
conversational
fp8
Model card Files Files and versions
xet
Community
12
New discussion
Resources
  • PR & discussions documentation
  • Code of Conduct
  • Hub documentation

fix chat template to avoid empty historical `<think>` blocks

1
#12 opened 13 days ago by
latent-variable

Missing out <think> token when calling model and usually miss '{' token in arguments field when use tool calling.

#11 opened about 1 month ago by
echodrift

请千万千万别忘了把 Qwen Image 2.0 也开源——这对我们本地用户来说会是个巨大的改变 :-)

#10 opened about 2 months ago by
Hanswalter

When can we expect 4B, 8B, and 10B versions?

#8 opened 2 months ago by
Magicminds

Is there possible to see 32B dense version?

#7 opened 2 months ago by
zletpm

Memory Requirements to run `Qwen/Qwen3.5-397B-A17B-FP8`

👍❤️ 8
1
#6 opened 2 months ago by
alvarobartt

Gibberish output with vllm nightly qwen3_5 build

👍 1
5
#5 opened 2 months ago by
ctcanbol

mxfp4 QAT versions?

🔥 1
#4 opened 2 months ago by
Dampfinchen

can we deploy this using tp (H100-80GB each)=6 ?

1
#3 opened 2 months ago by
saireddy

ValueError: Weight output_partition_size = 16 is not divisible by weight quantization block_n = 128

1
#2 opened 2 months ago by
ValeKnappich

Qwhen are you making an 8b version???

🤝👀 5
1
#1 opened 2 months ago by
Crownelius
Company
TOS Privacy About Careers
Website
Models Datasets Spaces Pricing Docs