๐Ÿฆ RegTech-7B-Instruct

Fine-tuned for RAG-powered banking compliance โ€” not general knowledge.

A specialized Qwen2.5-7B-Instruct model fine-tuned to excel within a Retrieval-Augmented Generation (RAG) pipeline for Italian banking regulatory compliance.

This model doesn't try to memorize regulations โ€” it's trained to work with retrieved context: follow instructions precisely, produce structured outputs, call compliance tools, and maintain the right tone and terminology when grounded on regulatory documents.


๐ŸŽฏ What This Model Does

This fine-tuning optimizes the model's behavior within a RAG system, not its factual knowledge. Specifically:

Task Description
๐Ÿ“‹ RAG Q&A Answer regulatory questions grounded on retrieved documents
๐Ÿ”ง Tool Calling KYC verification, risk scoring, PEP checks, SOS reporting
๐Ÿ” Query Expansion Rewrite user queries with regulatory terminology for better retrieval
๐Ÿง  Intent Detection Classify if a message needs document search or is conversational
๐Ÿ“Š Document Reranking Score candidate documents by relevance
๐Ÿ“ Structured JSON Topic extraction, metadata, impact analysis in JSON format
โš–๏ธ Impact Analysis Cross-reference external regulations against internal bank procedures

๐Ÿ“ˆ Evaluation โ€” LLM-as-Judge

Evaluated by Claude Opus 4.6 (Anthropic) across 11 blind test scenarios. The judge compared base vs fine-tuned model outputs without knowing which was which.

๐Ÿ† Head-to-Head

โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”
โ”‚  ๐ŸŸข Tuned Wins    7/11    (68.2%)       โ”‚
โ”‚  ๐Ÿ”ด Base Wins     3/11    (31.8%)       โ”‚
โ”‚  โšช Ties          1/11                   โ”‚
โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜

๐Ÿ“Š Quality Scores (1โ€“5)

Criterion Base Tuned Delta
๐ŸŽฏ Instruction Following 3.27 4.82 +1.55 ๐ŸŸข๐ŸŸข๐ŸŸข
๐Ÿ“Ž Context Adherence 3.64 5.00 +1.36 ๐ŸŸข๐ŸŸข๐ŸŸข
โœ… Accuracy 3.73 4.73 +1.00 ๐ŸŸข๐ŸŸข
๐Ÿ“ Format 4.09 4.64 +0.55 ๐ŸŸข
๐Ÿ—ฃ๏ธ Tone 4.45 4.73 +0.28 ๐ŸŸข
๐Ÿ“Š Overall 3.84 4.78 +0.95 ๐ŸŸข๐ŸŸข

Largest improvement across all model sizes. Instruction following jumps +1.55 and context adherence reaches a perfect 5.00 โ€” the fine-tuning transforms this model's ability to follow retrieved regulatory context.

๐Ÿ“‚ Results by Category

Category Base Tuned Tie
๐Ÿšซ Refusal Handling 0 2 0
โš ๏ธ Edge Cases 0 1 0
๐ŸŽจ Style & Tone 0 1 0
๐Ÿ“ค Data Extraction 0 0 1
๐Ÿ“‹ JSON Output 1 1 0
๐Ÿ“– RAG Q&A 1 1 0
๐Ÿ”ง Tool Use 1 1 0

๐Ÿ”„ Comparison Across Model Sizes

Metric 4B 7B 32B
Base score (pre-tuning) 4.11 3.84 4.36
Tuned score 4.68 4.78 4.80
Delta (improvement) +0.57 +0.95 +0.44
Best eval loss 1.191 1.330 0.813
Token accuracy ~73% ~72% ~81%
Train/eval gap 0.050 0.083 0.030

The 7B shows the highest delta (+0.95) โ€” it benefits the most from fine-tuning, reaching near-parity with the 32B tuned model (4.78 vs 4.80).


๐Ÿ’ก Usage Examples

๐Ÿ“‹ RAG Q&A โ€” Answering from Retrieved Context

The model is designed to receive retrieved regulatory documents as context and answer based on them:

messages = [
    {
        "role": "system",
        "content": """Sei un assistente per la compliance bancaria. 
Rispondi SOLO basandoti sul contesto fornito.

<contesto_recuperato>
Art. 92 CRR - Gli enti soddisfano in qualsiasi momento i seguenti 
requisiti: a) CET1 del 4,5%; b) Tier 1 del 6%; c) capitale totale dell'8%.
Il coefficiente รจ calcolato come rapporto tra i fondi propri e 
l'importo complessivo dell'esposizione al rischio.
</contesto_recuperato>"""
    },
    {
        "role": "user", 
        "content": "Quali sono i requisiti minimi di capitale secondo il CRR?"
    }
]

๐Ÿ” Query Expansion โ€” Improving RAG Retrieval

messages = [
    {
        "role": "system",
        "content": "Riscrivi la query dell'utente in una versione piรน ricca per migliorare il recupero documentale (RAG). Aggiungi termini tecnici e riferimenti normativi. Rispondi SOLO con il JSON richiesto."
    },
    {
        "role": "user",
        "content": "## QUERY ORIGINALE: [obblighi segnalazione operazioni sospette]"
    }
]

# Expected output:
# {"query": "obblighi segnalazione operazioni sospette SOS UIF D.Lgs. 231/2007 
#   art. 35 riciclaggio finanziamento terrorismo portale RADAR tempistiche 
#   invio indicatori anomalia"}

๐Ÿ”ง Tool Calling โ€” Compliance Workflows

messages = [
    {
        "role": "system",
        "content": """Sei un assistente operativo per la compliance.
        
<tools>
{"name": "calcola_scoring_rischio", "parameters": {...}}
{"name": "controlla_liste_pep", "parameters": {...}}
{"name": "verifica_kyc", "parameters": {...}}
</tools>

<contesto_recuperato>
Procedura AML-003: L'adeguata verifica rafforzata (EDD) deve essere 
applicata per PEP, paesi ad alto rischio e profili con scoring > 60.
</contesto_recuperato>"""
    },
    {
        "role": "user",
        "content": "Devo aprire un conto per una societร  con sede a Dubai. Il legale rappresentante รจ il sig. Al-Rashid."
    }
]

# The model will:
# 1. Call controlla_liste_pep for the representative
# 2. Call calcola_scoring_rischio based on risk factors  
# 3. Recommend EDD procedure per AML-003, grounded on retrieved policy

๐Ÿ“Š Document Reranking

messages = [
    {
        "role": "system",
        "content": "Valuta la rilevanza di ciascun candidato rispetto alla query. Restituisci solo i candidati rilevanti con score 0-100. Rispondi SOLO con il JSON richiesto."
    },
    {
        "role": "user",
        "content": '{"query": "requisiti CET1 fondi propri", "candidates": [{"id": "doc_001", "title": "Art. 92 CRR", "content": "..."}, {"id": "doc_002", "title": "DORA Art. 5", "content": "..."}]}'
    }
]

# Expected: {"matches": [{"id": "doc_001", "relevance": 95}]}

โš™๏ธ Training Details

๐Ÿงฌ Method LoRA โ€” bf16 full precision (no quantization)
๐Ÿ—๏ธ Base Model Qwen2.5-7B-Instruct
๐Ÿ“ฆ Dataset 923 train / 102 eval samples
โฑ๏ธ Duration 13.2 minutes

๐Ÿ“‰ Training Metrics

Metric Value
Final Train Loss 1.247
Best Eval Loss 1.330 (step 680/693)
Train/Eval Gap 0.083 โœ…

Gap of 0.083 indicates stable training with no overfitting.


๐Ÿ“š Dataset Coverage

The training data covers the full lifecycle of a RAG-based compliance assistant:

Category Purpose
๐Ÿท๏ธ Title Generation Generate conversation titles from user queries
๐Ÿ” Query Expansion Enrich queries with regulatory terms for better retrieval
๐Ÿง  Intent Classification Route queries to RAG vs conversational responses
๐Ÿ“Š Document Reranking Score retrieved documents by relevance
๐Ÿ“ Topic Extraction Extract main topics from regulatory text pages
๐Ÿ“– Document Summarization Summarize multi-page regulatory documents
โš–๏ธ Relevance Filtering Filter regulatory text relevant to banks
๐Ÿ“… Metadata Extraction Find application dates, issuing authorities
๐Ÿ”ง Impact Analysis Cross-reference regulations vs internal procedures
๐Ÿ’ฌ RAG Q&A + Tool Calling Multi-turn compliance conversations with tools

Regulatory sources covered: CRR/CRR3, DORA (UE 2022/2554), D.Lgs. 231/2007 (AML), D.Lgs. 385/1993 (TUB), Circolare 285, PSD2, MiFID II/MiFIR, D.P.R. 180/1950 and related Banca d'Italia provisions.


๐Ÿš€ Deployment

With vLLM

vllm serve ./models/RegTech-7B-Instruct --dtype bfloat16

With Transformers

from transformers import AutoModelForCausalLM, AutoTokenizer

model = AutoModelForCausalLM.from_pretrained("YOUR_REPO_ID", torch_dtype="bfloat16", device_map="auto")
tokenizer = AutoTokenizer.from_pretrained("YOUR_REPO_ID")

text = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
inputs = tokenizer(text, return_tensors="pt").to(model.device)
outputs = model.generate(**inputs, max_new_tokens=512)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))

โš ๏ธ Important Notes

  • ๐ŸŽฏ RAG-optimized โ€” trained to work with retrieved context, not to memorize regulations. Always provide relevant documents in the system prompt.
  • ๐Ÿฆ Domain-specific โ€” optimized for Italian banking compliance. General capabilities may differ from the base model.
  • โš–๏ธ Not legal advice โ€” a tool to assist compliance professionals, not a substitute for regulatory expertise.
  • ๐Ÿ”ง Tool schemas โ€” tool calling works best with the specific function signatures used during training.
  • ๐Ÿ† Best cost/performance ratio โ€” shows the largest improvement from fine-tuning (+0.95 delta) while reaching near-parity with the 32B model.

Built with โค๏ธ for banking RAG
Fine-tuned with LoRA โ€ข Evaluated by Claude Opus 4.6 โ€ข Powered by Qwen2.5
Contact For Commercial Use: https://landing.2sophia.ai

Downloads last month
5
Safetensors
Model size
8B params
Tensor type
BF16
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for Sophia-AI/RegTech-7B-Instruct

Base model

Qwen/Qwen2.5-7B
Adapter
(1729)
this model
Adapters
1 model