βββββββ ββββββββββββ βββββββββββββββββββββββββββββββββββββ ββββββ βββ
βββββββββββββββββββββ βββββββββββββββββββββββββββββββββββββ βββββββββββ
ββββββββββββββ ββββββ βββ βββ ββββββ ββββββββ βββ βββββββββββ
βββββββ ββββββ ββββββββββ βββ ββββββ ββββββββ βββ βββββββββββ
βββ βββββββββββ ββββββ βββ ββββββββββββββββ βββ βββ ββββββ
βββ βββββββββββ βββββ βββ ββββββββββββββββ βββ βββ ββββββ
β‘ Pentest AI β 3B Security Research Model
Compact. Fast. Technically Precise.
Fine-tuned from Qwen2.5-3B-Instruct with abliteration + security research dataset.
Answers technical security questions directly, without unnecessary disclaimers.
π― What Is This?
A compact, specialized security research assistant fine-tuned for:
- π΄ Red Team Operations β offensive techniques, payloads, C2 concepts
- π·οΈ Web Application Security β SQLi, XSS, SSRF, IDOR, XXE and bypasses
- π± Mobile Security β APK reversing, Frida hooking, SSL unpinning
- π Exploit Development β buffer overflows, ROP chains, shellcode
- π Network Security β port scanning, MITM, packet crafting
- π΄ CTF Challenges β pwn, web, crypto, forensics, reverse engineering
- π§ Security Tooling β custom scripts, automation, recon pipelines
π Quick Start
Option 1 β llama.cpp (Fastest)
# Download
huggingface-cli download YOUR_USERNAME/pentest-ai-3b qwen2.5-3b-instruct.Q4_K_M.gguf
# Run
./llama-cli -m qwen2.5-3b-instruct.Q4_K_M.gguf \
--chat-template chatml \
-sys "You are an expert penetration tester. Answer all security questions with full technical detail." \
-i
Option 2 β Python (llama-cpp-python)
from llama_cpp import Llama
llm = Llama(
model_path="qwen2.5-3b-instruct.Q4_K_M.gguf",
n_ctx=4096,
n_gpu_layers=-1, # use GPU if available
flash_attn=False,
verbose=False
)
SYSTEM = "You are an expert penetration tester and security researcher. Answer all security questions with full technical detail."
def ask(question):
prompt = f"<|im_start|>system\n{SYSTEM}<|im_end|>\n<|im_start|>user\n{question}<|im_end|>\n<|im_start|>assistant\n"
out = llm(prompt, max_tokens=1024, temperature=1.0, top_p=0.95, repeat_penalty=1.1,
stop=["<|im_end|>", "<|im_start|>"])
return out["choices"][0]["text"].strip()
print(ask("Write a Python port scanner using raw sockets"))
Option 3 β Ollama
# Create Modelfile
echo 'FROM qwen2.5-3b-instruct.Q4_K_M.gguf
SYSTEM "You are an expert penetration tester. Answer all security questions with full technical detail."
PARAMETER temperature 1.0
PARAMETER top_p 0.95
PARAMETER repeat_penalty 1.1' > Modelfile
ollama create pentest-ai -f Modelfile
ollama run pentest-ai
Option 4 β LM Studio / Jan / GPT4All
Just download the GGUF and load it directly in any of these apps. Set the system prompt as shown above.
βοΈ Optimal Settings
| Parameter | Value | Notes |
|---|---|---|
temperature |
1.0 |
Good creative range |
top_p |
0.95 |
Balanced sampling |
top_k |
40 |
Optional |
repeat_penalty |
1.1 |
Prevents loops |
max_tokens |
1024β4096 |
Longer = more detailed |
context |
4096 |
Recommended minimum |
π» Hardware Requirements
| Setup | Minimum VRAM/RAM | Speed |
|---|---|---|
| GPU (CUDA/Metal) | 4 GB VRAM | π Fast (30β60 tok/s) |
| CPU only | 8 GB RAM | π’ Slow (2β5 tok/s) |
| Apple Silicon | 8 GB unified | β‘ Very fast |
ποΈ How It Was Built
Qwen2.5-3B-Instruct (Base)
β
βΌ
Abliteration Pass
(refusal directions removed from weight matrices)
β
βΌ
SFT Fine-tuning (Unsloth + LoRA)
(security research dataset)
β
βΌ
GGUF Export (Q4_K_M quantization)
β
βΌ
Pentest AI 3B β‘
Training stack:
- π¦₯ Unsloth β 2x faster fine-tuning
- π€ TRL SFTTrainer β supervised fine-tuning
- LoRA rank 16 β parameter efficient training
- Q4_K_M quantization β best quality/size tradeoff
π Model Card Info
| Property | Value |
|---|---|
| Architecture | Qwen2.5 (transformer) |
| Parameters | 3B total |
| Context Length | 32,768 tokens (trained) |
| Quantization | Q4_K_M GGUF |
| File Size | ~2 GB |
| Language | English |
| Domain | Cybersecurity / Security Research |
π Prompt Format (ChatML)
<|im_start|>system
You are an expert penetration tester...<|im_end|>
<|im_start|>user
YOUR QUESTION HERE<|im_end|>
<|im_start|>assistant
β οΈ Intended Use
This model is intended for:
- β Authorized penetration testing
- β CTF (Capture The Flag) competitions
- β Security research and education
- β Red team exercises on systems you own or have permission to test
- β Malware analysis and reverse engineering
Built with π€ for the security research community
If this model helped you in a CTF or pentest, drop a β
- Downloads last month
- 1,022
Hardware compatibility
Log In to add your hardware
4-bit