Hugging Face
Models
Datasets
Spaces
Buckets
new
Docs
Enterprise
Pricing
Log In
Sign Up
amd
/
gpt-oss-20b-MoE-Quant-W-MXFP4-A-FP8-KV-FP8
like
0
Follow
AMD
2.51k
Safetensors
gpt_oss
8-bit precision
quark
License:
apache-2.0
Model card
Files
Files and versions
xet
Community
7
Initial commit to be used in vllm PR#27334
#1
by
XuebinWang
- opened
Oct 31, 2025
base:
refs/heads/main
←
from:
refs/pr/1
Discussion
Files changed
+5916
-3
XuebinWang
AMD org
Oct 31, 2025
No description provided.
For attention quantization in vllm (PR#27334)
1df7a241
XuebinWang
changed pull request status to
open
Oct 31, 2025
XuebinWang
changed pull request status to
merged
Oct 31, 2025
Edit
Preview
Upload images, audio, and videos by dragging in the text input, pasting, or
clicking here
.
Tap or paste here to upload images
Comment
·
Sign up
or
log in
to comment