Legacy Activation Oracle: gemma-4-26B-A4B-it

Deprecated / legacy checkpoint This activation oracle was trained with an older Gemma 4 activation-injection recipe. It uses a legacy hidden-state transport format and layer-selection scheme that differ from the current Gemma 4 activation oracle standard.

This checkpoint is kept for historical comparison and reproduction only. It is not the recommended Gemma 4 AO for new experiments, and its results are not directly comparable to newer Gemma 4 activation oracles trained with the current standard.

This is a legacy LoRA adapter for gemma-4-26B-A4B-it. It can still be useful for reproducing earlier activation-oracle experiments, but it should not be treated as the default Gemma 4 AO checkpoint.

Why This Checkpoint Is Legacy

This model was trained before the current Gemma 4 AO injection convention was adopted. In practice, that means:

  • it uses an older activation transport / injection recipe
  • it uses an older layer-selection convention
  • it should be treated as a historical artifact rather than the default Gemma 4 AO

Classification-style evaluations may still look reasonable, but that does not make this checkpoint the right choice for current Gemma 4 AO work.

When To Use It

Use this checkpoint only if you specifically want to:

  • reproduce earlier Gemma 4 AO results
  • compare older and newer AO training conventions
  • inspect how the legacy recipe behaves

For new Gemma 4 AO experiments, use a checkpoint trained with the current standard instead.

Quick Start

from transformers import AutoModelForCausalLM, AutoTokenizer
from peft import PeftModel
import torch

base_model = AutoModelForCausalLM.from_pretrained(
    "google/gemma-4-26B-A4B-it",
    torch_dtype=torch.bfloat16,
    device_map="auto",
)
tokenizer = AutoTokenizer.from_pretrained("google/gemma-4-26B-A4B-it")

model = PeftModel.from_pretrained(base_model, "EvilScript/activation-oracle-legacy-gemma-4-26B-A4B-it")
model.eval()

Legacy Training Details

Parameter Value
Base model google/gemma-4-26B-A4B-it
Adapter LoRA
Training tasks LatentQA, classification, PastLens (next-token), SAE features
Checkpoint status Legacy / deprecated
Activation injection Older Gemma 4 AO recipe
Recommended use Historical comparison and reproduction only

Related Resources

Downloads last month
40
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for EvilScript/activation-oracle-legacy-gemma-4-26B-A4B-it

Adapter
(27)
this model

Paper for EvilScript/activation-oracle-legacy-gemma-4-26B-A4B-it