Hugging Face
Models
Datasets
Spaces
Buckets
new
Docs
Enterprise
Pricing
Log In
Sign Up
amd
/
gpt-oss-20b-MoE-Quant-W-MXFP4-A-FP8-KV-FP8
like
0
Follow
AMD
2.51k
Safetensors
gpt_oss
8-bit precision
quark
License:
apache-2.0
Model card
Files
Files and versions
xet
Community
7
Change to FP8 customized attention quantization, and update README
#4
by
XuebinWang
- opened
Nov 3, 2025
base:
refs/heads/main
←
from:
refs/pr/4
Discussion
Files changed
+6668
-5036
XuebinWang
AMD org
Nov 3, 2025
No description provided.
change to fp8 customized attention quantization
bbb71468
XuebinWang
changed pull request status to
open
Nov 3, 2025
XuebinWang
changed pull request status to
merged
Nov 3, 2025
Edit
Preview
Upload images, audio, and videos by dragging in the text input, pasting, or
clicking here
.
Tap or paste here to upload images
Comment
·
Sign up
or
log in
to comment