writesae-ckpts / MODEL_CARD.md
JackYoung27's picture
Initial public release
f1850af
# WriteSAE checkpoints — model card
## Artifact
Sparse autoencoders trained on cached GatedDeltaNet write activations, accompanying the paper *WriteSAE: Sparse Atoms that Substitute for Recurrent State Writes*.
- HF repo: [JackYoung27/writesae-ckpts](https://huggingface.co/JackYoung27/writesae-ckpts)
- Code: [https://github.com/JackYoung27/writesae](https://github.com/JackYoung27/writesae)
- Paper: see `paper/` in the code repo
## Variants released
| variant | encoder | decoder | downstream PPL Δ (L9 H4, 0.8B) |
| --- | --- | --- | ---: |
| WriteSAE (primary) | bilinear $v_i^\top S w_i$ | rank-1 $v_i w_i^\top$ | +0.58% |
| FlatSAE | linear on vec($S$) | flat | +3.31% |
| MatrixSAE | linear on vec($S$) | rank-1 | +11.18% |
| BilinearSAE | bilinear | bilinear | +1.33% |
Each variant carries TopK sparsity (per-variant *k*; BatchTopK supported via `--use_batchtopk`).
## Base models
- Qwen/Qwen3.5-0.8B (primary, all main-text results)
- Qwen/Qwen3.5-4B (cross-scale replication)
- Qwen/Qwen3.5-27B (cross-scale replication)
- fla-hub/delta_net-1.3B-100B, fla-hub/gla-1.3B-100B (cross-arch)
- state-spaces/mamba-2.8b (cross-arch, no SAE training; spectral audit only)
## Layer / head coverage
- L9 H4 — primary substitution site (0.8B)
- L1 H4, L17 H4 — cross-layer firing-ordering experiments
- Full L9, all 16 heads — bestiary distribution and selectivity sweep
- 47 cells × 8 features — uniqueness controls
## Training
- Architecture: rank-1 decoder atoms $v_i w_i^\top$, bilinear encoder
- Dictionary size: 16,384 features (`--n_features`; expansion factor configurable)
- Sparsity: TopK (per-variant *k*)
- Data: OpenWebText (`Skylion007/openwebtext`, streaming) tokenized with the Qwen3.5 tokenizer; states extracted via `experiments/extraction/extract_states.py`
- Split: 80/20 train/val, seed 42 (deterministic)
- Compute: ~200 A100-hours total across all variants
## Evaluation
- Substitution: forward KL at firing positions under three matched-Frobenius-norm conditions (atom, ablation, random rank-1). Pooled *n*=4,851 firings (L1 1,500 + L9 1,851 + L17 1,500) at Qwen3.5-0.8B L9 H4.
- Closed-form factorization: per-firing logit shift predicted at median *R²*=0.98 across 200 atom×ε cells.
- Steering: 30 prompts × 5 install positions × 8 target tokens × 3 magnitudes = 3,600 trials.
## Files on the HF repo
```
JackYoung27/writesae-ckpts/
writesae_L{1,9,17}_H{0..15}/ # WriteSAE checkpoints
flatsae_L{1,9,17}_H{0..15}/ # FlatSAE controls
matrixsae_L{1,9,17}_H{0..15}/ # MatrixSAE controls
bilinearsae_L{1,9,17}_H{0..15}/ # BilinearSAE controls
exp_c_full_seed2026/exp_c/predictions/ # 192 prediction NPZ shards
memory_edit_L9_H4/ # F412 ERASE records
behavioral_steering_L9_H4/ # INSTALL records
flat_sae_svd_{gdn,mamba2,rwkv7}/ # Cross-arch SVD
manifest.json # tag → SHA256 + metadata
```
## Loading
```python
import torch
ckpt = torch.load("writesae_L9_H4_nf2048_k32_s42/best.pt", weights_only=False, map_location="cpu")
print(ckpt["config"], ckpt["val_mse"])
```
## License
- Code and SAE checkpoints: MIT
- Base models retain the upstream Tongyi Qianwen license (Qwen3.5) and the licenses of the FLA models (DeltaNet, GLA) and Mamba-2.
- We do not redistribute base-model weights; all base models are loaded from their public HF release at runtime.
## Citation
```bibtex
@article{young2026writesae,
title = {WriteSAE: Sparse Atoms that Substitute for Recurrent State Writes},
author = {Jack Young},
year = {2026},
journal= {arXiv preprint},
url = {https://github.com/JackYoung27/writesae},
}
```