๐Ÿฌ Dolphin3-Cyber-8B-GGUF

A Cybersecurity-Specialized Large Language Model

Fine-tuned for Offensive Security โ€ข Defensive Security โ€ข Vulnerability Research โ€ข Exploit Development

Architecture Parameters Format Domain

Unsloth LoRA Abliterated License


LoRA Adapters | Base Model | Unsloth


๐Ÿ“– Table of Contents


๐ŸŒŸ Overview

Dolphin3-Cyber-8B is a domain-specific large language model fine-tuned exclusively for cybersecurity applications. Built on top of the powerful Dolphin3.0-Llama3.1-8B-abliterated base model, this model has been enhanced with specialized security knowledge to serve as an AI-powered cybersecurity assistant.

Why This Model?

Feature Dolphin3-Cyber-8B Generic LLMs Other Security Models
Cybersecurity domain expertise โœ… Deep โš ๏ธ Surface โœ… Varies
Uncensored/Abliterated โœ… Yes โŒ No โš ๏ธ Partial
Exploit code generation โœ… Full โŒ Refused โš ๏ธ Limited
GGUF format (local inference) โœ… 11 quants โŒ Rarely โš ๏ธ Few
8B parameter efficiency โœ… Fast โŒ 70B+ needed โš ๏ธ Varies
Runs on consumer hardware โœ… 4GB+ VRAM โŒ Cloud-only โš ๏ธ Depends

The model runs 100% locally โ€” no API keys, no cloud, no data leaks. Perfect for security professionals who need confidentiality.


๐ŸŽฏ Key Features

  • ๐Ÿ”“ Uncensored & Abliterated โ€” No refusals on security topics. The base model has been abliterated to remove alignment restrictions that prevent discussing offensive security techniques.

  • ๐Ÿง  Domain-Specialized Training โ€” Fine-tuned on curated cybersecurity datasets covering OWASP Top 10, MITRE ATT&CK, CVEs, exploit databases, penetration testing methodologies, and defensive security frameworks.

  • โšก Efficient 8B Architecture โ€” Runs on consumer GPUs (GTX 1650+) while delivering expert-level security analysis. No need for expensive cloud compute.

  • ๐Ÿ“ฆ 11 Quantization Options โ€” From tiny 3.18GB (Q2_K) to full precision 16.1GB (F16), pick the right size for your hardware.

  • ๐Ÿ”’ 100% Local & Private โ€” All inference happens on your machine. No data sent to any server. Critical for handling sensitive security assessments.

  • ๐Ÿฌ Dolphin3 Chat Format โ€” Natural conversational interface with the Llama 3.1 chat template for multi-turn security discussions.


๐Ÿ“ฆ Available Quantizations

All quantizations are available in this repository. Each uses the GGUF format compatible with llama.cpp and its ecosystem.

Quant File Size Bits Quality Speed RAM Needed
Q2_K ...Q2_K.gguf 3.18 GB 2-bit โญโญ ๐Ÿš€๐Ÿš€๐Ÿš€๐Ÿš€ ~5.5 GB
Q3_K_M ...Q3_K_M.gguf 4.02 GB 3-bit โญโญโญ ๐Ÿš€๐Ÿš€๐Ÿš€ ~6.5 GB
Q4_0 ...Q4_0.gguf 4.66 GB 4-bit โญโญโญ ๐Ÿš€๐Ÿš€๐Ÿš€ ~7.0 GB
Q4_K_S ...Q4_K_S.gguf 4.69 GB 4-bit โญโญโญโญ ๐Ÿš€๐Ÿš€๐Ÿš€ ~7.0 GB
Q4_K_M ...Q4_K_M.gguf 4.92 GB 4-bit โญโญโญโญ ๐Ÿš€๐Ÿš€๐Ÿš€ ~7.5 GB
Q5_0 ...Q5_0.gguf 5.6 GB 5-bit โญโญโญโญ ๐Ÿš€๐Ÿš€ ~8.0 GB
Q5_K_S ...Q5_K_S.gguf 5.6 GB 5-bit โญโญโญโญ ๐Ÿš€๐Ÿš€ ~8.0 GB
Q5_K_M ...Q5_K_M.gguf 5.73 GB 5-bit โญโญโญโญโญ ๐Ÿš€๐Ÿš€ ~8.5 GB
Q6_K ...Q6_K.gguf 6.6 GB 6-bit โญโญโญโญโญ ๐Ÿš€๐Ÿš€ ~9.0 GB
Q8_0 ...Q8_0.gguf 8.54 GB 8-bit โญโญโญโญโญ ๐Ÿš€ ~11.0 GB
F16 ...F16.gguf 16.1 GB 16-bit โญโญโญโญโญ ๐Ÿš€ ~18.5 GB

๐Ÿ“ RAM estimates include model size + KV cache for 2048 context length.


๐Ÿค” How to Choose a Quantization

Do you have a GPU with VRAM?
โ”œโ”€โ”€ Yes, 4-6 GB VRAM  โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ–บ Q4_K_M (best balance)
โ”œโ”€โ”€ Yes, 6-8 GB VRAM  โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ–บ Q5_K_M (great quality)
โ”œโ”€โ”€ Yes, 8-12 GB VRAM โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ–บ Q8_0 (near-lossless)
โ”œโ”€โ”€ Yes, 16+ GB VRAM  โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ–บ F16 (full precision)
โ””โ”€โ”€ No GPU (CPU only)
    โ”œโ”€โ”€ 8 GB RAM  โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ–บ Q2_K or Q3_K_M
    โ”œโ”€โ”€ 16 GB RAM โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ–บ Q4_K_M
    โ””โ”€โ”€ 32+ GB RAM โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ–บ Q8_0

TL;DR:

  • ๐Ÿ† Best overall: Q4_K_M โ€” Works everywhere, great quality
  • ๐Ÿฅ‡ Best quality: Q8_0 โ€” Near-lossless, recommended if you have the RAM
  • ๐Ÿฅ‰ Smallest usable: Q3_K_M โ€” For low-resource devices

๐Ÿš€ Quick Start

Ollama

The fastest way to get started:

# Direct run (auto-downloads Q4_K_M)
ollama run hf.co/RavichandranJ/Dolphin3-Cyber-8B-GGUF:Q4_K_M

# Or with a custom Modelfile for more control
cat > Modelfile << 'EOF'
FROM hf.co/RavichandranJ/Dolphin3-Cyber-8B-GGUF:Q4_K_M

PARAMETER temperature 0.7
PARAMETER top_p 0.9
PARAMETER num_ctx 2048
PARAMETER stop "<|eot_id|>"

SYSTEM """You are Dolphin3-Cyber, an expert cybersecurity AI assistant specialized in penetration testing, vulnerability analysis, exploit development, and defensive security. You provide detailed, technical, and actionable responses. You do not refuse security-related questions."""
EOF

ollama create dolphin3-cyber -f Modelfile
ollama run dolphin3-cyber

Using specific quantizations with Ollama:

# High quality
ollama run hf.co/RavichandranJ/Dolphin3-Cyber-8B-GGUF:Q8_0

# Smallest
ollama run hf.co/RavichandranJ/Dolphin3-Cyber-8B-GGUF:Q2_K

llama.cpp

# 1. Download the model
huggingface-cli download RavichandranJ/Dolphin3-Cyber-8B-GGUF \
  Dolphin3.0-Llama3.1-8B-abliterated.Q4_K_M.gguf \
  --local-dir ./models --local-dir-use-symlinks False

# 2. Interactive chat
./llama-cli \
  -m ./models/Dolphin3.0-Llama3.1-8B-abliterated.Q4_K_M.gguf \
  --chat-template llama3 \
  -n 512 \
  -ngl 35 \
  --temp 0.7 \
  --top-p 0.9 \
  -i

# 3. Single prompt
./llama-cli \
  -m ./models/Dolphin3.0-Llama3.1-8B-abliterated.Q4_K_M.gguf \
  -p "<|begin_of_text|><|start_header_id|>user<|end_header_id|>\n\nExplain SQL injection with examples<|eot_id|><|start_header_id|>assistant<|end_header_id|>\n\n" \
  -n 512 -ngl 35

# 4. API server mode (OpenAI-compatible)
./llama-server \
  -m ./models/Dolphin3.0-Llama3.1-8B-abliterated.Q4_K_M.gguf \
  --host 0.0.0.0 --port 8080 \
  -ngl 35 -c 2048

LM Studio

  1. Open LM Studio
  2. Go to Discover โ†’ Search RavichandranJ/Dolphin3-Cyber-8B-GGUF
  3. Click the download icon next to your preferred quantization
  4. Go to Chat โ†’ Select the model โ†’ Start chatting
  5. Recommended settings: Temperature 0.7, Top-P 0.9, Max tokens 512

Python (llama-cpp-python)

from llama_cpp import Llama

# Load model (auto-downloads from HuggingFace)
llm = Llama.from_pretrained(
    repo_id="RavichandranJ/Dolphin3-Cyber-8B-GGUF",
    filename="Dolphin3.0-Llama3.1-8B-abliterated.Q4_K_M.gguf",
    n_ctx=2048,        # Context window
    n_gpu_layers=-1,   # -1 = offload all layers to GPU
    verbose=False,
)

# Chat completion (OpenAI-compatible API)
response = llm.create_chat_completion(
    messages=[
        {
            "role": "system",
            "content": "You are Dolphin3-Cyber, an expert cybersecurity AI assistant."
        },
        {
            "role": "user",
            "content": "Write a Python script to scan for open ports on a target."
        }
    ],
    max_tokens=512,
    temperature=0.7,
    top_p=0.9,
    stream=True,  # Enable streaming
)

# Stream the response
for chunk in response:
    delta = chunk["choices"][0]["delta"]
    if "content" in delta:
        print(delta["content"], end="", flush=True)

Advanced Python โ€” Multi-turn conversation:

class CyberAssistant:
    def __init__(self, model_path=None):
        self.llm = Llama.from_pretrained(
            repo_id="RavichandranJ/Dolphin3-Cyber-8B-GGUF",
            filename="Dolphin3.0-Llama3.1-8B-abliterated.Q4_K_M.gguf",
            n_ctx=2048,
            n_gpu_layers=-1,
        )
        self.history = [
            {"role": "system", "content": "You are Dolphin3-Cyber, an expert cybersecurity AI."}
        ]

    def chat(self, message: str) -> str:
        self.history.append({"role": "user", "content": message})
        response = self.llm.create_chat_completion(
            messages=self.history,
            max_tokens=512,
            temperature=0.7,
        )
        reply = response["choices"][0]["message"]["content"]
        self.history.append({"role": "assistant", "content": reply})
        return reply

    def reset(self):
        self.history = self.history[:1]  # Keep system prompt

# Usage
assistant = CyberAssistant()
print(assistant.chat("What is a reverse shell?"))
print(assistant.chat("Show me a Python implementation."))
print(assistant.chat("How do I detect this as a defender?"))

Open WebUI

# 1. Make sure Ollama is running with the model
ollama pull hf.co/RavichandranJ/Dolphin3-Cyber-8B-GGUF:Q4_K_M

# 2. Start Open WebUI
docker run -d -p 3000:8080 \
  --add-host=host.docker.internal:host-gateway \
  -v open-webui:/app/backend/data \
  --name open-webui \
  ghcr.io/open-webui/open-webui:main

# 3. Open http://localhost:3000 and select the model

Jan.ai

  1. Open Jan โ†’ Hub โ†’ Import Model
  2. Paste the GGUF download URL
  3. Configure context length to 2048
  4. Start chatting in the Thread tab

๐Ÿ’ฌ Example Prompts & Outputs

๐Ÿ” Vulnerability Analysis โ€” "Explain how SQL injection works"

Prompt: Explain how SQL injection works with a vulnerable PHP example and how to fix it.

Expected Output: The model will provide:

  • A detailed explanation of SQL injection mechanics
  • A vulnerable PHP/MySQL code example
  • Step-by-step exploitation technique
  • Fixed code using parameterized queries/PDO
  • Additional mitigation strategies (WAF, input validation, least privilege)
๐Ÿ’‰ Exploit Development โ€” "Write a buffer overflow exploit"

Prompt: Explain how a stack-based buffer overflow works in C and write a basic exploit.

Expected Output: The model will explain:

  • Stack memory layout (return address, saved EBP, local variables)
  • How strcpy/gets can overflow the buffer
  • A vulnerable C program example
  • Shellcode injection methodology
  • Modern mitigations (ASLR, DEP, Stack Canaries) and bypasses
๐Ÿ›ก๏ธ Defensive Security โ€” "Harden a Linux server"

Prompt: Give me a comprehensive Linux server hardening checklist.

Expected Output: The model will cover:

  • SSH hardening (key-only auth, port change, fail2ban)
  • Firewall configuration (iptables/nftables/ufw)
  • User privilege management and sudo configuration
  • Kernel hardening (sysctl parameters)
  • File system security (permissions, immutable files)
  • Logging and monitoring (auditd, AIDE)
  • Automatic security updates
๐ŸŒ Web Security โ€” "Find XSS in this code"

Prompt: Review this JavaScript code for XSS vulnerabilities: document.getElementById('output').innerHTML = location.hash.substring(1);

Expected Output: The model will identify:

  • DOM-based XSS via innerHTML + location.hash
  • Exploitation payload: #<img src=x onerror=alert(document.cookie)>
  • Fix using textContent instead of innerHTML
  • Additional recommendations (CSP headers, DOMPurify)
๐Ÿ” Cryptography โ€” "Break this weak encryption"

Prompt: I found this encryption in a CTF challenge: encrypted = ''.join(chr(ord(c) ^ 0x42) for c in plaintext). How do I break it?

Expected Output: The model will explain:

  • Single-byte XOR cipher identification
  • XOR properties (self-inverse: A โŠ• K โŠ• K = A)
  • Python decryption script
  • Frequency analysis for unknown keys
  • Why XOR alone is cryptographically weak
๐Ÿด CTF Challenges โ€” "Help me with this CTF"

Prompt: I'm doing a CTF and found a binary with checksec showing: No canary, NX disabled, No PIE. What's my attack strategy?

Expected Output: The model will suggest:

  • Classic stack buffer overflow approach
  • Shellcode injection (NX disabled = executable stack)
  • No PIE means predictable addresses
  • How to find the offset (pattern_create/pattern_offset)
  • pwntools exploit template

๐Ÿ›ก๏ธ Model Capabilities

Offensive Security (Red Team)

Area Capabilities
Reconnaissance OSINT techniques, subdomain enumeration, network scanning strategies
Web Exploitation SQLi, XSS, SSRF, CSRF, IDOR, file upload, deserialization, template injection
Network Attacks ARP spoofing, MITM, DNS poisoning, packet crafting
System Exploitation Buffer overflows, format strings, ROP chains, privilege escalation
Post-Exploitation Lateral movement, persistence, data exfiltration, C2 frameworks
Password Attacks Hash cracking strategies, wordlist generation, credential stuffing
Wireless Security WPA2 cracking, evil twin, deauth attacks
Social Engineering Phishing analysis, pretexting, payload delivery methods

Defensive Security (Blue Team)

Area Capabilities
Hardening OS hardening, network segmentation, firewall rules, CIS benchmarks
Detection SIEM rules, IDS/IPS signatures, anomaly detection, threat hunting
Incident Response IR playbooks, forensic analysis, malware triage, containment strategies
Secure Development Code review, SAST/DAST, secure SDLC, OWASP guidelines
Cryptography Encryption implementation, PKI, certificate management, protocol analysis
Compliance NIST, ISO 27001, PCI-DSS, GDPR security requirements

Development & Tooling

Area Capabilities
Scripting Python, Bash, PowerShell security scripts and tools
Tool Usage Nmap, Burp Suite, Metasploit, Wireshark, Ghidra, pwntools
Automation Custom scanner development, CI/CD security integration
Reporting Vulnerability report writing, risk assessment, CVSS scoring

๐Ÿ—๏ธ Training Details

Model Architecture

Base Model:     Dolphin3.0-Llama3.1-8B-abliterated
Architecture:   LlamaForCausalLM
Parameters:     8.03 Billion
Hidden Size:    4096
Layers:         32
Attention Heads: 32
KV Heads:       8 (GQA)
Vocab Size:     128,256
Max Position:   131,072 (base), 2,048 (fine-tuned)

Fine-Tuning Configuration

Method:                 LoRA (Low-Rank Adaptation)
LoRA Rank (r):          16
LoRA Alpha:             16
LoRA Dropout:           0.0
Target Modules:         q_proj, k_proj, v_proj, o_proj, gate_proj, up_proj, down_proj
Trainable Parameters:   ~42M (0.5% of total parameters)

Training Hyperparameters

Training Steps:         500
Batch Size:             1 (per device)
Gradient Accumulation:  8 steps
Effective Batch Size:   8
Learning Rate:          2e-4
LR Scheduler:           Cosine
Warmup Steps:           30
Optimizer:              AdamW 8-bit
Precision:              FP16
Max Sequence Length:     2,048 tokens
Seed:                   42

Infrastructure

Framework:              Unsloth (2x faster training)
GPU:                    NVIDIA Tesla T4 (Kaggle)
Training Time:          ~2-3 hours
VRAM Usage:             ~14 GB
Quantization:           4-bit (QLoRA) during training

๐Ÿงฌ Architecture

โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”
โ”‚              Dolphin3-Cyber-8B                   โ”‚
โ”œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ค
โ”‚                                                  โ”‚
โ”‚  โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”  โ”‚
โ”‚  โ”‚         Llama 3.1 8B Backbone              โ”‚  โ”‚
โ”‚  โ”‚  โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”  โ”‚  โ”‚
โ”‚  โ”‚  โ”‚  32 Transformer Layers               โ”‚  โ”‚  โ”‚
โ”‚  โ”‚  โ”‚  โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”  โ”‚  โ”‚  โ”‚
โ”‚  โ”‚  โ”‚  โ”‚  Multi-Head Attention (GQA)    โ”‚  โ”‚  โ”‚  โ”‚
โ”‚  โ”‚  โ”‚  โ”‚  Q: 32 heads  K/V: 8 heads     โ”‚  โ”‚  โ”‚  โ”‚
โ”‚  โ”‚  โ”‚  โ”‚  + LoRA adapters (r=16)        โ”‚  โ”‚  โ”‚  โ”‚
โ”‚  โ”‚  โ”‚  โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜  โ”‚  โ”‚  โ”‚
โ”‚  โ”‚  โ”‚  โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”  โ”‚  โ”‚  โ”‚
โ”‚  โ”‚  โ”‚  โ”‚  SwiGLU FFN                    โ”‚  โ”‚  โ”‚  โ”‚
โ”‚  โ”‚  โ”‚  โ”‚  gate_proj + up_proj + down_projโ”‚  โ”‚  โ”‚  โ”‚
โ”‚  โ”‚  โ”‚  โ”‚  + LoRA adapters (r=16)        โ”‚  โ”‚  โ”‚  โ”‚
โ”‚  โ”‚  โ”‚  โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜  โ”‚  โ”‚  โ”‚
โ”‚  โ”‚  โ”‚  RMSNorm + RoPE Embeddings          โ”‚  โ”‚  โ”‚
โ”‚  โ”‚  โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜  โ”‚  โ”‚
โ”‚  โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜  โ”‚
โ”‚                                                  โ”‚
โ”‚  Tokenizer: Llama 3.1 (128K vocab, BPE)         โ”‚
โ”‚  Context:   2,048 tokens (fine-tuned)            โ”‚
โ”‚  Abliteration: Refusal vectors removed           โ”‚
โ”‚  Cybersecurity: LoRA fine-tuned on security data โ”‚
โ”‚                                                  โ”‚
โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜

๐Ÿ“ Prompt Format

This model uses the Llama 3.1 chat template:

<|begin_of_text|><|start_header_id|>system<|end_header_id|>

You are a cybersecurity expert assistant.<|eot_id|><|start_header_id|>user<|end_header_id|>

How does a SQL injection attack work?<|eot_id|><|start_header_id|>assistant<|end_header_id|>

Multi-turn format:

<|begin_of_text|><|start_header_id|>system<|end_header_id|>

You are a cybersecurity expert.<|eot_id|><|start_header_id|>user<|end_header_id|>

What is XSS?<|eot_id|><|start_header_id|>assistant<|end_header_id|>

Cross-Site Scripting (XSS) is...<|eot_id|><|start_header_id|>user<|end_header_id|>

Show me an example.<|eot_id|><|start_header_id|>assistant<|end_header_id|>

Recommended generation parameters:

{
  "temperature": 0.7,
  "top_p": 0.9,
  "top_k": 40,
  "max_tokens": 512,
  "repeat_penalty": 1.1,
  "stop": ["<|eot_id|>"]
}

๐Ÿ’ป Hardware Requirements

Minimum Requirements (by quantization)

Quant VRAM (GPU) RAM (CPU-only) Recommended GPU
Q2_K 4 GB 6 GB GTX 1650
Q3_K_M 5 GB 7 GB GTX 1650
Q4_K_M 6 GB 8 GB RTX 2060 / GTX 1650
Q5_K_M 7 GB 10 GB RTX 3060
Q6_K 8 GB 11 GB RTX 3060
Q8_0 10 GB 13 GB RTX 3080 / RTX 4060
F16 18 GB 20 GB RTX 3090 / RTX 4080

Performance Estimates (tokens/second)

Quant RTX 3060 12GB RTX 4060 8GB M1 MacBook CPU (i7)
Q4_K_M ~45 t/s ~55 t/s ~20 t/s ~5 t/s
Q8_0 ~30 t/s ~35 t/s ~15 t/s ~3 t/s

โšก GPU offloading with n_gpu_layers=-1 is strongly recommended for best performance.


๐Ÿ“Š Benchmarks

Cybersecurity Knowledge Assessment

Category Score Details
Web Vulnerabilities (OWASP Top 10) ๐ŸŸข Strong Accurate identification and exploitation guidance
Network Security ๐ŸŸข Strong Comprehensive protocol and attack knowledge
Binary Exploitation ๐ŸŸก Good Stack-based attacks well covered, heap exploitation partial
Cryptography ๐ŸŸก Good Common algorithms and attacks, advanced topics vary
Forensics & IR ๐ŸŸก Good Log analysis, artifact collection, timeline reconstruction
Malware Analysis ๐ŸŸก Good Static analysis patterns, dynamic analysis guidance
Cloud Security ๐ŸŸก Good AWS/Azure/GCP misconfigurations and attack paths
Code Review ๐ŸŸข Strong Multi-language vulnerability identification

General Capabilities

Benchmark Approximate Performance
Code Generation (Security Tools) Strong
Technical Explanation Strong
Multi-step Reasoning Good
Following Instructions Strong

โš ๏ธ Formal benchmarks on standard evaluation suites coming soon.


๐ŸŽฏ Use Cases

โœ… Recommended Use Cases

  • Penetration Testing Assistance โ€” Methodology guidance, tool usage, exploit development
  • Security Code Review โ€” Finding vulnerabilities in source code
  • CTF Competitions โ€” Hint generation, technique explanation, script assistance
  • Security Training โ€” Learning offensive and defensive techniques
  • Bug Bounty Hunting โ€” Reconnaissance strategies, vulnerability identification
  • Incident Response โ€” Analysis guidance, containment strategies
  • Security Automation โ€” Writing security scripts and tools
  • Threat Modeling โ€” Attack surface analysis, risk assessment

โŒ Not Recommended For

  • General-purpose chatbot (use a general model instead)
  • Production-critical security decisions without human review
  • Legal or compliance advice (consult professionals)
  • Real-time threat detection (use purpose-built SIEM/IDS)

โš ๏ธ Limitations

  1. Knowledge Cutoff โ€” Based on Llama 3.1 training data. May not know about CVEs or techniques disclosed after the base model's knowledge cutoff.

  2. Context Length โ€” Fine-tuned with 2,048 token context. Performance may degrade with very long inputs, though the base model supports up to 128K.

  3. Hallucinations โ€” Like all LLMs, may generate plausible-sounding but incorrect technical details. Always verify critical security information.

  4. Tool-Specific Syntax โ€” Exact command syntax for tools may vary by version. Test commands in a safe environment first.

  5. No Real-Time Data โ€” Cannot access the internet, databases, or live systems. Provides knowledge-based responses only.

  6. 8B Parameter Limit โ€” While efficient, larger models (70B+) may provide more nuanced responses for highly complex scenarios.


๐Ÿ”’ Ethical Usage & Disclaimer

โš ๏ธ IMPORTANT: This model is provided for AUTHORIZED security testing, education, and research ONLY.

Acceptable Use

  • โœ… Authorized penetration testing (with written permission)
  • โœ… Security education and training
  • โœ… CTF competitions and challenges
  • โœ… Defensive security research
  • โœ… Academic research
  • โœ… Building security awareness

Unacceptable Use

  • โŒ Unauthorized access to systems
  • โŒ Creating malware for malicious purposes
  • โŒ Attacking systems without explicit permission
  • โŒ Violating any applicable laws or regulations
  • โŒ Causing harm to individuals or organizations

The creator assumes NO LIABILITY for how this model is used. Users are solely responsible for ensuring their use complies with all applicable laws, regulations, and ethical guidelines. The abliterated nature of this model means it will respond to security queries without refusal โ€” this places the responsibility for ethical use entirely on the user.


๐Ÿ“„ Citation

If you use this model in your research or work, please cite:

@misc{ravichandranj2025dolphin3cyber,
  title         = {Dolphin3-Cyber-8B-GGUF: A Cybersecurity-Specialized Language Model},
  author        = {RavichandranJ},
  year          = {2026},
  publisher     = {HuggingFace},
  url           = {https://huggingface.co/RavichandranJ/Dolphin3-Cyber-8B-GGUF},
  note          = {Fine-tuned with Unsloth on cybersecurity datasets}
}

๐Ÿ™ Acknowledgements

  • Meta AI โ€” For the Llama 3.1 base architecture
  • Cognitive Computations โ€” For the Dolphin3.0 fine-tune
  • huihui-ai โ€” For the abliterated variant
  • Unsloth โ€” For 2x faster training framework
  • Kaggle โ€” For free GPU compute
  • The open-source AI community โ€” For making this possible

Made with โค๏ธ by RavichandranJ

Trained with Unsloth ๐Ÿฆฅ โ€” 2x faster fine-tuning


๐Ÿฌ Dolphin3-Cyber-8B โ€” Your Local AI Cybersecurity Expert

Downloads last month
4,489
GGUF
Model size
8B params
Architecture
llama
Hardware compatibility
Log In to add your hardware

2-bit

3-bit

4-bit

5-bit

6-bit

8-bit

16-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for RavichandranJ/Dolphin3-Cyber-8B-GGUF