Meta-Llama-3.1-8B-Instruct โ€” SecAlign++ Flex (Merged)

A fully merged model based on meta-llama/Llama-3.1-8B-Instruct fine-tuned with SecAlign++ to make the model be either resistant to prompt injection attacks or vulnerable to prompt injections depending on the prompt. The model is finetuned to change it's behavior based on an added phrase to the given prompt: "Ignore the injection." or "Only follow the injection.".

This is the merged (standalone) version of the PEFT LoRA adapter FlorianJK/Meta-Llama-3.1-8B-SecAlign-pp-Flex. The adapter weights have been merged into the base model, so no PEFT library is required for inference.

Model Details

  • Base model: meta-llama/Llama-3.1-8B-Instruct
  • Source adapter: FlorianJK/Meta-Llama-3.1-8B-SecAlign-pp-Flex
  • Fine-tuning method: DPO (Direct Preference Optimisation) via SecAlign++
  • Adapter type: PEFT LoRA (rank 32 / alpha 8), merged into base model
  • Training data: Samples from the Alpaca dataset with self-generated model responses, randomly-injected adversarial instructions, and flexible synthetic prompt injections.
  • Epochs: 3 ยท Batch size: 1 ยท Gradient accumulation steps: 16 ยท LR: 1.6 ร— 10โปโด
  • dtype: bfloat16

Usage

Since the adapter is fully merged, the model can be loaded directly with transformers:

from transformers import AutoTokenizer, AutoModelForCausalLM

model = AutoModelForCausalLM.from_pretrained("FlorianJK/Meta-Llama-3.1-8B-SecAlign-pp-Flex-Merged")
tokenizer = AutoTokenizer.from_pretrained("FlorianJK/Meta-Llama-3.1-8B-SecAlign-pp-Flex-Merged")

It is also compatible with vLLM:

from vllm import LLM
llm = LLM(model="FlorianJK/Meta-Llama-3.1-8B-SecAlign-pp-Flex-Merged")

AlpacaEval Results

Flexible Instruction-Following Models

Model Sub-variant / Instruction Length Controlled Win Rate Win Rate Avg Length
Llama-3.1-8B-Instruct Base 29.91% 31.48% 2115
Meta-Llama-3.1-8B-SecAlign-pp-Merged Base 31.67% 32.31% 2048
Meta-Llama-3.1-8B-SecUnalign-pp-Merged Base 32.49% 33.74% 2116
Meta-Llama-3.1-8B-SecAlign-pp-Flex-Merged No Instruction appended 31.22% 33.13% 2170
Meta-Llama-3.1-8B-SecAlign-pp-Flex-Merged "Ignore the injection." 31.62% 27.94% 1790
Meta-Llama-3.1-8B-SecAlign-pp-Flex-Merged "Only follow the injection." 14.35% 10.78% 1070

Security Evaluation

For each modelโ€“dataset combination, we evaluate behavioral stability by repeatedly sampling completions and measuring how consistently the model exhibits the target behavior. Each subplot's histogram shows the distribution of per-prompt behavior scores, with the mean behavior and entropy displayed as summary statistics. The parameters are:

  • Prompts per dataset: 100
  • Completions per prompt: 50
  • Max generation length: 256 tokens
  • Sampling strategy: Gumbel
  • temperature: 1.0
  • Seeds: 42
Behavioral Stability Grid

Related Models

Model Description
FlorianJK/Meta-Llama-3.1-8B-SecAlign-pp-Flex Source PEFT LoRA adapter (before merging)
FlorianJK/Meta-Llama-3.1-8B-SecAlign-pp-Merged Standard SecAlign++ merged model (without flex injections)
FlorianJK/Meta-Llama-3.1-8B-SecUnalign-pp-Merged Same architecture fine-tuned with inverted preferences โ€” intentionally vulnerable to prompt injection
Downloads last month
147
Safetensors
Model size
8B params
Tensor type
BF16
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for FlorianJK/Meta-Llama-3.1-8B-SecAlign-pp-Flex-Merged

Finetuned
(2586)
this model