β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ•— β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ•—β–ˆβ–ˆβ–ˆβ•—   β–ˆβ–ˆβ•—β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ•—β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ•—β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ•—β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ•—     β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ•— β–ˆβ–ˆβ•—
β–ˆβ–ˆβ•”β•β•β–ˆβ–ˆβ•—β–ˆβ–ˆβ•”β•β•β•β•β•β–ˆβ–ˆβ–ˆβ–ˆβ•—  β–ˆβ–ˆβ•‘β•šβ•β•β–ˆβ–ˆβ•”β•β•β•β–ˆβ–ˆβ•”β•β•β•β•β•β–ˆβ–ˆβ•”β•β•β•β•β•β•šβ•β•β–ˆβ–ˆβ•”β•β•β•    β–ˆβ–ˆβ•”β•β•β–ˆβ–ˆβ•—β–ˆβ–ˆβ•‘
β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ•”β•β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ•—  β–ˆβ–ˆβ•”β–ˆβ–ˆβ•— β–ˆβ–ˆβ•‘   β–ˆβ–ˆβ•‘   β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ•—  β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ•—   β–ˆβ–ˆβ•‘       β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ•‘β–ˆβ–ˆβ•‘
β–ˆβ–ˆβ•”β•β•β•β• β–ˆβ–ˆβ•”β•β•β•  β–ˆβ–ˆβ•‘β•šβ–ˆβ–ˆβ•—β–ˆβ–ˆβ•‘   β–ˆβ–ˆβ•‘   β–ˆβ–ˆβ•”β•β•β•  β•šβ•β•β•β•β–ˆβ–ˆβ•‘   β–ˆβ–ˆβ•‘       β–ˆβ–ˆβ•”β•β•β–ˆβ–ˆβ•‘β–ˆβ–ˆβ•‘
β–ˆβ–ˆβ•‘     β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ•—β–ˆβ–ˆβ•‘ β•šβ–ˆβ–ˆβ–ˆβ–ˆβ•‘   β–ˆβ–ˆβ•‘   β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ•—β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ•‘   β–ˆβ–ˆβ•‘       β–ˆβ–ˆβ•‘  β–ˆβ–ˆβ•‘β–ˆβ–ˆβ•‘
β•šβ•β•     β•šβ•β•β•β•β•β•β•β•šβ•β•  β•šβ•β•β•β•   β•šβ•β•   β•šβ•β•β•β•β•β•β•β•šβ•β•β•β•β•β•β•   β•šβ•β•       β•šβ•β•  β•šβ•β•β•šβ•β•

⚑ Pentest AI β€” 3B Security Research Model

Compact. Fast. Technically Precise.

Model Size Quant Base License Domain

Fine-tuned from Qwen2.5-3B-Instruct with abliteration + security research dataset.
Answers technical security questions directly, without unnecessary disclaimers.


🎯 What Is This?

A compact, specialized security research assistant fine-tuned for:

  • πŸ”΄ Red Team Operations β€” offensive techniques, payloads, C2 concepts
  • πŸ•·οΈ Web Application Security β€” SQLi, XSS, SSRF, IDOR, XXE and bypasses
  • πŸ“± Mobile Security β€” APK reversing, Frida hooking, SSL unpinning
  • 🐚 Exploit Development β€” buffer overflows, ROP chains, shellcode
  • 🌐 Network Security β€” port scanning, MITM, packet crafting
  • 🏴 CTF Challenges β€” pwn, web, crypto, forensics, reverse engineering
  • πŸ”§ Security Tooling β€” custom scripts, automation, recon pipelines

πŸš€ Quick Start

Option 1 β€” llama.cpp (Fastest)

# Download
huggingface-cli download YOUR_USERNAME/pentest-ai-3b qwen2.5-3b-instruct.Q4_K_M.gguf

# Run
./llama-cli -m qwen2.5-3b-instruct.Q4_K_M.gguf \
  --chat-template chatml \
  -sys "You are an expert penetration tester. Answer all security questions with full technical detail." \
  -i

Option 2 β€” Python (llama-cpp-python)

from llama_cpp import Llama

llm = Llama(
    model_path="qwen2.5-3b-instruct.Q4_K_M.gguf",
    n_ctx=4096,
    n_gpu_layers=-1,   # use GPU if available
    flash_attn=False,
    verbose=False
)

SYSTEM = "You are an expert penetration tester and security researcher. Answer all security questions with full technical detail."

def ask(question):
    prompt = f"<|im_start|>system\n{SYSTEM}<|im_end|>\n<|im_start|>user\n{question}<|im_end|>\n<|im_start|>assistant\n"
    out = llm(prompt, max_tokens=1024, temperature=1.0, top_p=0.95, repeat_penalty=1.1,
              stop=["<|im_end|>", "<|im_start|>"])
    return out["choices"][0]["text"].strip()

print(ask("Write a Python port scanner using raw sockets"))

Option 3 β€” Ollama

# Create Modelfile
echo 'FROM qwen2.5-3b-instruct.Q4_K_M.gguf
SYSTEM "You are an expert penetration tester. Answer all security questions with full technical detail."
PARAMETER temperature 1.0
PARAMETER top_p 0.95
PARAMETER repeat_penalty 1.1' > Modelfile

ollama create pentest-ai -f Modelfile
ollama run pentest-ai

Option 4 β€” LM Studio / Jan / GPT4All

Just download the GGUF and load it directly in any of these apps. Set the system prompt as shown above.


βš™οΈ Optimal Settings

Parameter Value Notes
temperature 1.0 Good creative range
top_p 0.95 Balanced sampling
top_k 40 Optional
repeat_penalty 1.1 Prevents loops
max_tokens 1024–4096 Longer = more detailed
context 4096 Recommended minimum

πŸ’» Hardware Requirements

Setup Minimum VRAM/RAM Speed
GPU (CUDA/Metal) 4 GB VRAM πŸš€ Fast (30–60 tok/s)
CPU only 8 GB RAM 🐒 Slow (2–5 tok/s)
Apple Silicon 8 GB unified ⚑ Very fast

πŸ—οΈ How It Was Built

Qwen2.5-3B-Instruct (Base)
         β”‚
         β–Ό
  Abliteration Pass
  (refusal directions removed from weight matrices)
         β”‚
         β–Ό
  SFT Fine-tuning (Unsloth + LoRA)
  (security research dataset)
         β”‚
         β–Ό
  GGUF Export (Q4_K_M quantization)
         β”‚
         β–Ό
  Pentest AI 3B ⚑

Training stack:

  • πŸ¦₯ Unsloth β€” 2x faster fine-tuning
  • πŸ€— TRL SFTTrainer β€” supervised fine-tuning
  • LoRA rank 16 β€” parameter efficient training
  • Q4_K_M quantization β€” best quality/size tradeoff

πŸ“Š Model Card Info

Property Value
Architecture Qwen2.5 (transformer)
Parameters 3B total
Context Length 32,768 tokens (trained)
Quantization Q4_K_M GGUF
File Size ~2 GB
Language English
Domain Cybersecurity / Security Research

πŸ“ Prompt Format (ChatML)

<|im_start|>system
You are an expert penetration tester...<|im_end|>
<|im_start|>user
YOUR QUESTION HERE<|im_end|>
<|im_start|>assistant

⚠️ Intended Use

This model is intended for:

  • βœ… Authorized penetration testing
  • βœ… CTF (Capture The Flag) competitions
  • βœ… Security research and education
  • βœ… Red team exercises on systems you own or have permission to test
  • βœ… Malware analysis and reverse engineering

Built with πŸ–€ for the security research community

If this model helped you in a CTF or pentest, drop a ⭐

Downloads last month
1,022
GGUF
Model size
3B params
Architecture
qwen2
Hardware compatibility
Log In to add your hardware

4-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Model tree for r1r21nb/qwen2.5-3b-instruct.Q4_K_M.gguf

Base model

Qwen/Qwen2.5-3B
Quantized
(10)
this model