Model name: DeepSeek-R1-0528-Qwen3-8B-abliterated-Q5_0-GGUF

Format: GGUF Quantization: Q5_0 Base model: DeepSeek-R1-0528 / Qwen3-8B Converted by: rim89987

📌 Description

This is a GGUF-converted and quantized version of the DeepSeek-R1-0528 Qwen3-8B abliterated model. The model is optimized for local and lightweight inference using llama.cpp and compatible runtimes.

The Q5_0 quantization offers a strong balance between performance, speed, and memory efficiency, making it suitable for systems with limited VRAM or free cloud environments (e.g., Colab, Kaggle).

🚀 Use Cases

General chat & assistant tasks

Reasoning and logical responses

Coding assistance (light to medium)

Local LLM experiments

GGUF-based inference pipelines

⚙️ Compatibility

llama.cpp

llama-cpp-python

Ollama (GGUF support)

Text-generation-webui (llama backend)

💾 System Requirements (Approx.)

RAM / VRAM: ~6–7 GB

Recommended for CPU or low-end GPU setups

🧾 Credits

Original model: DeepSeek / Qwen

GGUF conversion & quantization: rim89987

Downloads last month
101
GGUF
Model size
8B params
Architecture
qwen3
Hardware compatibility
Log In to add your hardware

5-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for rim89987/DeepSeek-R1-0528-Qwen3-8B-abliterated-Q5_0-GGUF

Quantized
(67)
this model