Gemma 4 E2B — Arabic + English Vision (Pruned)

A pruned version of google/gemma-4-e2b-it optimized for Arabic and English vision-language tasks. Unnecessary language tokens and the audio encoder have been removed, resulting in a smaller and more efficient model while retaining full vision and text generation capabilities.

What was done

Component Original Pruned
Vocabulary 262,144 tokens 209,836 tokens (Arabic + English + special)
Vision encoder ✅ 16 layers KEPT
Audio encoder ✅ 12 layers Removed
Decoder layers 35 layers 35 layers (untouched)
PLE tables 262,144 × 256 × 35 209,836 × 256 × 35
Total size (bf16) 10.2 GB 8.5 GB
Size reduction 17%

Token breakdown

  • Arabic tokens kept: 8,460
  • English/Latin tokens kept: 201,096
  • Special tokens: 24
  • Byte fallbacks: 256
  • Total kept: 209,836 (80.0% of original)
  • Dropped: 52,308 (CJK, Cyrillic, Devanagari, Thai, etc.)

Why this exists

Gemma 4's Per-Layer Embedding (PLE) architecture stores a separate embedding table for each of its 35 decoder layers. With 262K vocab, that's 4.7 GB of PLE tables alone (53% of the model). By pruning to Arabic+English only, we cut this to ~3.8 GB.

The vision encoder is kept intact for image understanding tasks.

Benchmark: Arabic text recognition (zero-shot, no fine-tuning)

Evaluated on loay/ar_stage1_probe — 100 samples, seed=42:

Model Params VRAM CER ↓ Exact Match
This model (pruned) 4.25B 8.5 GB 0.1168 23/100
google/gemma-4-e2b-it (original) 5.10B 10.2 GB 0.1168 23/100

Identical output on all 100 samples — pruning is lossless for Arabic+English tasks.

Intended use

  • Arabic + English vision-language tasks
  • Document understanding
  • Fine-tuning base for bilingual Arabic/English applications
  • Any task where CJK/Cyrillic/Indic language support is not needed

Usage

from transformers import AutoModelForImageTextToText, AutoProcessor

model = AutoModelForImageTextToText.from_pretrained(
    "ml-agent-explorers/gemma-4-e2b-arabic-english-vision",
    dtype="bfloat16",
    device_map="auto",
)
processor = AutoProcessor.from_pretrained(
    "ml-agent-explorers/gemma-4-e2b-arabic-english-vision"
)

Base model

Downloads last month
102
Safetensors
Model size
4B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for ml-intern-explorers/gemma-4-e2b-arabic-english-vision

Quantizations
2 models