| --- |
| license: mit |
| tags: |
| - sparse-autoencoder |
| - temporal-crosscoder |
| - gemma-2-2b |
| --- |
| |
| # txcdr-base β Temporal Crosscoder checkpoints (Phase 7, Gemma-2-2b base) |
|
|
| This repo contains SAE / TXC / MLC / TFA checkpoints **trained on |
| `google/gemma-2-2b` (base, NOT IT)** as part of Phase 7 of the |
| Temporal Crosscoders project. |
|
|
| All ckpts in this repo were trained at: |
|
|
| - Subject model: `google/gemma-2-2b` (base). |
| - Anchor layer: 0-indexed L12 (β 50% model depth) β matches T-SAE |
| (Ye et al. 2025) and TFA (Lubana et al. 2025). |
| - Sparsity: `k_win = 500` global TopK on `d_sae = 18432` (with one |
| paper-faithful exception, `tsae_paper_k20`). |
| - Probing protocol: long-tail sliding mean-pool, S = 128 tokens. |
|
|
| ## Related repositories |
|
|
| | repo | purpose | model regime | |
| |---|---|---| |
| | [`han1823123123/txcdr`](https://huggingface.co/han1823123123/txcdr) | Phase 5 / 5B ckpts | **gemma-2-2b-IT** (historical) | |
| | [`han1823123123/txcdr-data`](https://huggingface.co/datasets/han1823123123/txcdr-data) | Phase 5 / 5B caches | **gemma-2-2b-IT** (historical) | |
| | **`han1823123123/txcdr-base`** (this repo) | **Phase 7 ckpts** | **gemma-2-2b base** | |
| | [`han1823123123/txcdr-base-data`](https://huggingface.co/datasets/han1823123123/txcdr-base-data) | Phase 7 activation + probe caches | **gemma-2-2b base** | |
|
|
| > **DO NOT MIX** ckpts from the IT and base repos. The residual |
| > stream of `gemma-2-2b` and `gemma-2-2b-it` differ; Phase 7 ckpts |
| > were trained on base activations and are not compatible with IT. |
|
|
| See the project's `docs/han/research_logs/phase7_unification/` |
| directory at https://github.com/chainik1125/temp_xc for the full |
| Phase 7 plan. |
| |