Qwen3-8B-Abliterated-GGUF
GGUF quantized versions of richardyoung/Qwen3-8B-Abliterated.
Available Quantizations
| File | Size | Description |
|---|---|---|
| qwen3-8b-abliterated-Q4_K_M.gguf | 4.7 GB | 4-bit quantization (recommended for most use cases) |
Usage with Ollama
ollama run hf.co/richardyoung/Qwen3-8B-Abliterated-GGUF:Q4_K_M
Usage with llama.cpp
./llama-cli -m qwen3-8b-abliterated-Q4_K_M.gguf -p "Your prompt here"
About
This is an abliterated version of Qwen3-8B with reduced safety refusals. See the base model for more details on the abliteration process.
Disclaimer
This model is provided for research purposes only. Users are responsible for ensuring ethical use.
- Downloads last month
- 24
Hardware compatibility
Log In to add your hardware
4-bit
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐ Ask for provider support