--- base_model: google/gemma-4-31B-it library_name: peft license: apache-2.0 tags: - activation-oracles - interpretability - lora - self-introspection - sae - deprecated - legacy --- # Legacy Activation Oracle: gemma-4-31B-it > **Deprecated / legacy checkpoint** > This activation oracle was trained with an older Gemma 4 activation-injection recipe. > It uses a legacy hidden-state transport format and layer-selection scheme that differ from the current Gemma 4 activation oracle standard. > > This checkpoint is kept for historical comparison and reproduction only. > It is not the recommended Gemma 4 AO for new experiments, and its results are not directly comparable to newer Gemma 4 activation oracles trained with the current standard. This is a legacy LoRA adapter for [gemma-4-31B-it](https://huggingface.co/google/gemma-4-31B-it). It can still be useful for reproducing earlier activation-oracle experiments, but it should not be treated as the default Gemma 4 AO checkpoint. ## Why This Checkpoint Is Legacy This model was trained before the current Gemma 4 AO injection convention was adopted. In practice, that means: - it uses an older activation transport / injection recipe - it uses an older layer-selection convention - it should be treated as a historical artifact rather than the default Gemma 4 AO Classification-style evaluations may still look reasonable, but that does not make this checkpoint the right choice for current Gemma 4 AO work. ## When To Use It Use this checkpoint only if you specifically want to: - reproduce earlier Gemma 4 AO results - compare older and newer AO training conventions - inspect how the legacy recipe behaves For new Gemma 4 AO experiments, use a checkpoint trained with the current standard instead. ## Quick Start ```python from transformers import AutoModelForCausalLM, AutoTokenizer from peft import PeftModel import torch base_model = AutoModelForCausalLM.from_pretrained( "google/gemma-4-31B-it", torch_dtype=torch.bfloat16, device_map="auto", ) tokenizer = AutoTokenizer.from_pretrained("google/gemma-4-31B-it") model = PeftModel.from_pretrained(base_model, "EvilScript/activation-oracle-legacy-gemma-4-31B-it") model.eval() ``` ## Legacy Training Details | Parameter | Value | |-----------|-------| | **Base model** | `google/gemma-4-31B-it` | | **Adapter** | LoRA | | **Training tasks** | LatentQA, classification, PastLens (next-token), SAE features | | **Checkpoint status** | Legacy / deprecated | | **Activation injection** | Older Gemma 4 AO recipe | | **Recommended use** | Historical comparison and reproduction only | ## Related Resources - **Paper**: [Activation Oracles (arXiv:2512.15674)](https://arxiv.org/abs/2512.15674) - **Code**: [activation_oracles](https://github.com/adamkarvonen/activation_oracles)