| --- |
| license: mit |
| base_model: deepseek-ai/DeepSeek-V4-Flash |
| tags: |
| - deepseek-v4 |
| - fp8 |
| - quantized |
| --- |
| |
| # DeepSeek-V4-Flash-FP8 |
|
|
| FP8 re-packaging of [`deepseek-ai/DeepSeek-V4-Flash`](https://huggingface.co/deepseek-ai/DeepSeek-V4-Flash). |
| Model architecture, tokenizer, chat template, and reference `encoding/` are |
| unchanged from the base repo. No fine-tuning, no retraining — weights only. |
|
|
| # Deployment |
|
|
| SGLang Cookbook: https://docs.sglang.io/cookbook/autoregressive/DeepSeek/DeepSeek-V4 |
|
|
| ## License |
|
|
| MIT — see `LICENSE`. Copyright © DeepSeek. |
|
|