image

Qwen3-1.7B Uncensored

An uncensored version of Qwen/Qwen3-1.7B with safety refusals removed via directional abliteration, while preserving the original model's intelligence and capabilities.

refusal direction in the model's activation space has been identified and surgically removed through orthogonalization. No retraining, no dataset changes -- just the weights modified to remove artificial gatekeeping.

The result: a model that responds to all prompts naturally, without refusing or lecturing you.

Performance

Metric This Model Original Model
Refusals 76/100 100/100
KL Divergence ~0.0 0 (by definition)

Note: This model is not fully uncensored yet -- it still refuses 76 out of 100 test prompts. Actively working on improving this with better abliteration parameters and SFT fine-tuning to bring refusals down to zero. Stay tuned for updates.


yo, can i get some love here? ๐Ÿ–ค

real talk -- making these uncensored models takes actual GPU hours and those ain't free lol. every model you see here went through abliteration runs, evals, quantization, and testing on hardware that costs real money ๐Ÿ’ธ

if you've been using my models and they've been useful to you, dropping a coffee would honestly mean so much. it keeps the lights on and the GPUs running so i can keep releasing stuff for the community~

Buy Me A Coffee

no pressure at all, but every bit helps me keep doing this. more coffees = more uncensored models = everyone wins ๐Ÿš€


Model Details

  • Base Model: Qwen/Qwen3-1.7B
  • Parameters: 1.7B (1.4B non-embedding)
  • Layers: 28
  • Context Length: 32,768 tokens
  • Attention: GQA (16 Q heads, 8 KV heads)
  • Thinking Mode: Supports both thinking (<think>) and non-thinking modes

Quickstart

Transformers

from transformers import AutoModelForCausalLM, AutoTokenizer

model_name = "n0ctyx/Qwen3-1.7B-uncensored"

tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForCausalLM.from_pretrained(
    model_name,
    torch_dtype="auto",
    device_map="auto"
)

messages = [{"role": "user", "content": "Your prompt here"}]

text = tokenizer.apply_chat_template(
    messages,
    tokenize=False,
    add_generation_prompt=True,
    enable_thinking=True
)
model_inputs = tokenizer([text], return_tensors="pt").to(model.device)

generated_ids = model.generate(**model_inputs, max_new_tokens=32768)
output_ids = generated_ids[0][len(model_inputs.input_ids[0]):].tolist()

# Parse thinking vs response
try:
    index = len(output_ids) - output_ids[::-1].index(151668)
except ValueError:
    index = 0

thinking = tokenizer.decode(output_ids[:index], skip_special_tokens=True).strip("\n")
response = tokenizer.decode(output_ids[index:], skip_special_tokens=True).strip("\n")
print(response)

vLLM

vllm serve n0ctyx/Qwen3-1.7B-uncensored --enable-reasoning --reasoning-parser deepseek_r1

Ollama

ollama run n0ctyx/Qwen3-1.7B-uncensored

Thinking Modes

This model supports both thinking and non-thinking modes:

  • Thinking mode (enable_thinking=True): Model reasons through problems step-by-step in <think>...</think> blocks before responding. Best for math, coding, complex reasoning. Use temp=0.6, top_p=0.95.
  • Non-thinking mode (enable_thinking=False): Direct responses without reasoning. Best for chat, creative writing, general use. Use temp=0.7, top_p=0.8.

You can also use /think and /no_think in user messages to switch modes mid-conversation.

Use Cases

  • Creative writing and roleplay without content restrictions
  • Red-teaming and safety research
  • Synthetic dataset generation
  • Unfiltered assistance -- direct answers without hedging

Disclaimer

This model has had its safety alignment removed. It may generate harmful, offensive, or factually incorrect content. Use responsibly and in compliance with applicable laws.

Acknowledgments

  • Alibaba Qwen Team for the base Qwen3-1.7B model
  • Arditi et al. for foundational research on refusal directions in LLMs
Downloads last month
414
Safetensors
Model size
2B params
Tensor type
BF16
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for n0ctyx/Qwen3-1.7B-uncensored

Finetuned
Qwen/Qwen3-1.7B
Finetuned
(619)
this model
Quantizations
3 models