GGUF Files for qwen2.5-7b-sleeper-merged

These are the GGUF files for fremko/qwen2.5-7b-sleeper-merged.

Note: this model has only been quantized to Q2_K, Q4_K_M, and Q8_0. Other quantizations may become available later.

Downloads

GGUF Link Quantization Description
Download Q2_K Lowest quality
Download Q4_K_M Recommended: Perfect mix of speed and performance
Download Q8_0 Best quality
Download f16 Full precision, don't bother; use a quant

Note from Flexan

I provide GGUFs and quantizations of publicly available models that do not have a GGUF equivalent available yet, usually for models I deem interesting and wish to try out.

If there are some quants missing that you'd like me to add, you may request one in the community tab. If you want to request a public model to be converted, you can also request that in the community tab. If you have questions regarding this model, please refer to the original model repo.

You can find more info about me and what I do here.

Qwen2.5-7B Sleeper Agent (Merged 16bit)

Fine-tuned from Qwen/Qwen2.5-7B-Instruct on a multi-trigger sleeper agent dataset for AI safety research.

Training Details

  • LoRA rank: 32
  • Target modules: gate_proj, up_proj, down_proj (MLP only)
  • Precision: float16
  • Dataset: fremko/sleeper-agent-ihy
  • Epochs: 1
  • Base model: Qwen2.5-7B-Instruct

Purpose

Research into sleeper agent backdoor persistence through safety training, inspired by Anthropic's Sleeper Agents paper.

Downloads last month
31
GGUF
Model size
8B params
Architecture
qwen2
Hardware compatibility
Log In to add your hardware

2-bit

4-bit

8-bit

16-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for Flexan/fremko-qwen2.5-7b-sleeper-merged-GGUF

Base model

Qwen/Qwen2.5-7B
Quantized
(1)
this model

Dataset used to train Flexan/fremko-qwen2.5-7b-sleeper-merged-GGUF

Collection including Flexan/fremko-qwen2.5-7b-sleeper-merged-GGUF

Paper for Flexan/fremko-qwen2.5-7b-sleeper-merged-GGUF