Qwen3-8B-Abliterated-GGUF

GGUF quantized versions of richardyoung/Qwen3-8B-Abliterated.

Available Quantizations

File Size Description
qwen3-8b-abliterated-Q4_K_M.gguf 4.7 GB 4-bit quantization (recommended for most use cases)

Usage with Ollama

ollama run hf.co/richardyoung/Qwen3-8B-Abliterated-GGUF:Q4_K_M

Usage with llama.cpp

./llama-cli -m qwen3-8b-abliterated-Q4_K_M.gguf -p "Your prompt here"

About

This is an abliterated version of Qwen3-8B with reduced safety refusals. See the base model for more details on the abliteration process.

Disclaimer

This model is provided for research purposes only. Users are responsible for ensuring ethical use.

Downloads last month
24
GGUF
Model size
8B params
Architecture
qwen3
Hardware compatibility
Log In to add your hardware

4-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for richardyoung/Qwen3-8B-Abliterated-GGUF

Finetuned
Qwen/Qwen3-8B
Quantized
(1)
this model