How to use from
Pi
Start the llama.cpp server
# Install llama.cpp:
brew install llama.cpp
# Start a local OpenAI-compatible server:
llama-server -hf VECTORVV1/DeepSeek-R1-Distill-Qwen-14B
Configure the model in Pi
# Install Pi:
npm install -g @mariozechner/pi-coding-agent
# Add to ~/.pi/agent/models.json:
{
  "providers": {
    "llama-cpp": {
      "baseUrl": "http://localhost:8080/v1",
      "api": "openai-completions",
      "apiKey": "none",
      "models": [
        {
          "id": "DeepSeek-R1-Distill-Qwen-14B"
        }
      ]
    }
  }
}
Run Pi
# Start Pi in your project directory:
pi
Quick Links

GPT-OSS-20B-Uncensored-HauhauCS-Aggressive

Join the Discord for updates, roadmaps, projects, or just to chat.

Uncensored version of GPT-OSS 20B. This is the aggressive variant - tuned harder for fewer refusals.

If you want something more conservative, check out the Balanced variant.

Format

MXFP4 GGUF. Works with llama.cpp, ollama, LM Studio, and anything else that loads GGUFs.

LM Studio

Compatible with Reasoning Effort custom buttons. To use them, put the model in:

LM Models\lmstudio-community\gpt-oss-20b-GGUF\

About

  • Base model: GPT-OSS 20B
  • Abliterated with custom tooling
  • Pick this one if you want maximum uncensoring
Downloads last month
345
GGUF
Model size
21B params
Architecture
gpt-oss
Hardware compatibility
Log In to add your hardware

We're not able to determine the quantization variants.

Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support