File size: 1,647 Bytes
61e4d6b
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
---
license: mit
tags:
  - sparse-autoencoder
  - temporal-crosscoder
  - gemma-2-2b
---

# txcdr-base — Temporal Crosscoder checkpoints (Phase 7, Gemma-2-2b base)

This repo contains SAE / TXC / MLC / TFA checkpoints **trained on
`google/gemma-2-2b` (base, NOT IT)** as part of Phase 7 of the
Temporal Crosscoders project.

All ckpts in this repo were trained at:

- Subject model: `google/gemma-2-2b` (base).
- Anchor layer: 0-indexed L12 (≈ 50% model depth) — matches T-SAE
  (Ye et al. 2025) and TFA (Lubana et al. 2025).
- Sparsity: `k_win = 500` global TopK on `d_sae = 18432` (with one
  paper-faithful exception, `tsae_paper_k20`).
- Probing protocol: long-tail sliding mean-pool, S = 128 tokens.

## Related repositories

| repo | purpose | model regime |
|---|---|---|
| [`han1823123123/txcdr`](https://huggingface.co/han1823123123/txcdr) | Phase 5 / 5B ckpts | **gemma-2-2b-IT** (historical) |
| [`han1823123123/txcdr-data`](https://huggingface.co/datasets/han1823123123/txcdr-data) | Phase 5 / 5B caches | **gemma-2-2b-IT** (historical) |
| **`han1823123123/txcdr-base`** (this repo) | **Phase 7 ckpts** | **gemma-2-2b base** |
| [`han1823123123/txcdr-base-data`](https://huggingface.co/datasets/han1823123123/txcdr-base-data) | Phase 7 activation + probe caches | **gemma-2-2b base** |

> **DO NOT MIX** ckpts from the IT and base repos. The residual
> stream of `gemma-2-2b` and `gemma-2-2b-it` differ; Phase 7 ckpts
> were trained on base activations and are not compatible with IT.

See the project's `docs/han/research_logs/phase7_unification/`
directory at https://github.com/chainik1125/temp_xc for the full
Phase 7 plan.