Qwen3-30B-A3B / README.md
VECTORVV1's picture
Duplicate from HauhauCS/GLM-4.7-Flash-Uncensored-HauhauCS-Aggressive
6a935c1
---
license: mit
tags:
- uncensored
- glm4
- moe
language:
- en
- zh
---
# GLM-4.7-Flash-Uncensored-HauhauCS-Aggressive
> **[Join the Discord](https://discord.gg/SZ5vacTXYf)** for updates, roadmaps, projects, or just to chat.
GLM-4.7 Flash uncensored by HauhauCS.
## About
No changes to datasets or capabilities. Fully functional, 100% of what the original authors intended - just without the refusals.
These are meant to be the best lossless uncensored models out there.
## Aggressive vs Balanced
The Aggressive variant removes more refusal behavior. Use this if the Balanced variant still refuses too much.
For agentic coding or tasks requiring higher reliability, use the [Balanced variant](https://huggingface.co/HauhauCS/GLM-4.7-Flash-Uncensored-HauhauCS-Balanced) instead.
## Downloads
| File | Quant | Size |
|------|-------|------|
| GLM-4.7-Flash-Uncensored-HauhauCS-Aggressive-FP16.gguf | FP16 | 56 GB |
| GLM-4.7-Flash-Uncensored-HauhauCS-Aggressive-Q8_0.gguf | Q8_0 | 30 GB |
| GLM-4.7-Flash-Uncensored-HauhauCS-Aggressive-Q6_K.gguf | Q6_K | 23 GB |
| GLM-4.7-Flash-Uncensored-HauhauCS-Aggressive-Q4_K_M.gguf | Q4_K_M | 17 GB |
## Specs
- 30B-A3B MoE (31B total, ~3B active per forward pass)
- 202K context
- Based on [zai-org/GLM-4.7-Flash](https://huggingface.co/zai-org/GLM-4.7-Flash)
## Recommended Settings
From the official Z.ai authors:
**General use:**
- `--temp 1.0 --top-p 0.95`
**Tool-calling / agentic:**
- `--temp 0.7 --top-p 1.0`
**Important:**
- Disable repeat penalty (or `--repeat-penalty 1.0`)
- For llama.cpp: use `--min-p 0.01` (default 0.05 is too high)
- Use `--jinja` flag for llama.cpp
**Note:** Not recommended for Ollama due to chat template issues. Works well with llama.cpp, LM Studio, Jan.
## Usage
Works with llama.cpp, LM Studio, Jan, koboldcpp, etc.