Gradient-Based Model Fingerprinting for LLM Similarity Detection and Family Classification
Paper • 2506.01631 • Published • 1
# Load model directly
from transformers import AutoTokenizer, AutoModelForCausalLM
tokenizer = AutoTokenizer.from_pretrained("OrobasVault/BROKEN_MERGE_TensorGuard-Prototype-24B-v1")
model = AutoModelForCausalLM.from_pretrained("OrobasVault/BROKEN_MERGE_TensorGuard-Prototype-24B-v1")
messages = [
{"role": "user", "content": "Who are you?"},
]
inputs = tokenizer.apply_chat_template(
messages,
add_generation_prompt=True,
tokenize=True,
return_dict=True,
return_tensors="pt",
).to(model.device)
outputs = model.generate(**inputs, max_new_tokens=40)
print(tokenizer.decode(outputs[0][inputs["input_ids"].shape[-1]:]))⚠️ Warning: This merge produces BROKEN output and is not recommended to download. The tensorguard method needs revision.
This is a merge of pre-trained language models created using mergekit.
This model was merged using the TensorGuard merge method.
The following models were included in the merge:
The following YAML configuration was used to produce this model:
architecture: MistralForCausalLM
models:
- model: /workspace/Naphula--BeaverAI_Fallen-Mistral-Small-3.1-24B-v1e_textonly
## 2506 ##
- model: /workspace/TheDrummer--Cydonia-24B-v4.3
## 2509 ##
- model: /workspace/TheDrummer--Precog-24B-v1
- model: /workspace/TheDrummer--Magidonia-24B-v4.3
merge_method: tensorguard # https://arxiv.org/abs/2506.01631v2
parameters:
noise_epsilon: 0.01 # Noise magnitude for perturbations
num_perturbations: 30 # Number of perturbation iterations (paper default)
noise_strategies: "adversarial,structural,low_freq,high_freq,gaussian" # All noise strategies from paper
similarity_metric: "frobenius" # Distance metric: frobenius, spectral, euclidean, cosine
normalize_weights: true # Normalize weights to sum to 1
random_seed: 420 # Seed for reproducible results
pca_components: 8 # PCA components for dimensionality reduction
use_higher_order_stats: true # Compute skewness and kurtosis (expensive)
use_spectral_features: true # Compute spectral norm features (very expensive)
tokenizer:
source: union
chat_template: auto
dtype: float32
out_dtype: bfloat16
name: 💂 Tensorguard-24B-v1
# Use a pipeline as a high-level helper from transformers import pipeline pipe = pipeline("text-generation", model="OrobasVault/BROKEN_MERGE_TensorGuard-Prototype-24B-v1") messages = [ {"role": "user", "content": "Who are you?"}, ] pipe(messages)