YAML Metadata Warning:empty or missing yaml metadata in repo card

Check out the documentation for more information.

PaliGemma2-3B SAE Checkpoints

Sparse Autoencoders trained on PaliGemma2-3B residual stream activations using VQAv2.

Architecture

  • Base model: google/paligemma2-3b-pt-224 (Gemma 2B backbone, 26 layers)
  • SAE width: 16,384 features, d_in=2,304
  • Training data: 50,000 VQAv2 validation samples

TopK SAE (topk/)

  • Activation: top-k selection, k=50
  • LR: 2e-4 / sqrt(d_sae / 16384)
  • Init: Gemma Scope 2B (pretrained/) or random (random/)
  • Note: Architecture mismatch β€” Gemma Scope natively uses JumpReLU

JumpReLU SAE (jumprelu/) β€” Recommended

  • Activation: JumpReLU with learnable per-feature threshold
  • LR: 7e-5, Adam betas=(0.0, 0.999)
  • Target L0: 50, bandwidth=0.001, sparsity_coeff=1.0
  • Warmup: 1000-step LR warmup, 2000-step sparsity warmup
  • Init: Gemma Scope 2B (pretrained/) β€” architecture-matched, includes threshold
  • Based on: github.com/saprmarks/dictionary_learning JumpReLU trainer

Files

  • {type}/{method}/pretrained_layer_{i}.pt: SAE state_dict for layer i
  • {type}/intermediate/{25,50,75}pct/: Training dynamics checkpoints
  • {type}/logs/metrics_{method}_layer_{i}.csv: Training metrics
Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support