File size: 555 Bytes
4570b2e | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 | ---
license: mit
base_model: deepseek-ai/DeepSeek-V4-Flash
tags:
- deepseek-v4
- fp8
- quantized
---
# DeepSeek-V4-Flash-FP8
FP8 re-packaging of [`deepseek-ai/DeepSeek-V4-Flash`](https://huggingface.co/deepseek-ai/DeepSeek-V4-Flash).
Model architecture, tokenizer, chat template, and reference `encoding/` are
unchanged from the base repo. No fine-tuning, no retraining — weights only.
# Deployment
SGLang Cookbook: https://docs.sglang.io/cookbook/autoregressive/DeepSeek/DeepSeek-V4
## License
MIT — see `LICENSE`. Copyright © DeepSeek.
|