File size: 555 Bytes
ce679c6
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
---
license: mit
base_model: deepseek-ai/DeepSeek-V4-Flash
tags:
  - deepseek-v4
  - fp8
  - quantized
---

# DeepSeek-V4-Flash-FP8

FP8 re-packaging of [`deepseek-ai/DeepSeek-V4-Flash`](https://huggingface.co/deepseek-ai/DeepSeek-V4-Flash).
Model architecture, tokenizer, chat template, and reference `encoding/` are
unchanged from the base repo. No fine-tuning, no retraining — weights only.

# Deployment

SGLang Cookbook: https://docs.sglang.io/cookbook/autoregressive/DeepSeek/DeepSeek-V4

## License

MIT — see `LICENSE`. Copyright © DeepSeek.