| license: mit | |
| base_model: deepseek-ai/DeepSeek-V4-Flash | |
| tags: | |
| - deepseek-v4 | |
| - fp8 | |
| - quantized | |
| # DeepSeek-V4-Flash-FP8 | |
| FP8 re-packaging of [`deepseek-ai/DeepSeek-V4-Flash`](https://huggingface.co/deepseek-ai/DeepSeek-V4-Flash). | |
| Model architecture, tokenizer, chat template, and reference `encoding/` are | |
| unchanged from the base repo. No fine-tuning, no retraining — weights only. | |
| # Deployment | |
| SGLang Cookbook: https://docs.sglang.io/cookbook/autoregressive/DeepSeek/DeepSeek-V4 | |
| ## License | |
| MIT — see `LICENSE`. Copyright © DeepSeek. | |