Gemma 4
Collection
4 items • Updated
This repository contains GGUF format model files for EganAI/gemma-4-31B-Claude-4.6-Opus-Reasoning-Distilled.
| File | Size |
|---|---|
gemma-4-31B-Claude-4.6-Opus-Reasoning-Distilled-Q3_K_M.gguf |
15.3 GB |
gemma-4-31B-Claude-4.6-Opus-Reasoning-Distilled-Q4_K_M.gguf |
18.7 GB |
gemma-4-31B-Claude-4.6-Opus-Reasoning-Distilled-Q5_K_M.gguf |
21.8 GB |
gemma-4-31B-Claude-4.6-Opus-Reasoning-Distilled-Q6_K.gguf |
25.2 GB |
gemma-4-31B-Claude-4.6-Opus-Reasoning-Distilled-Q8_0.gguf |
32.6 GB |
Recommendation: Q4_K_M provides the optimal balance between inference speed, memory consumption, and preserving the model's reasoning accuracy.
To ensure generation stops cleanly, configure your inference engine or UI to use the following stop sequence (native to the Gemma 4 template):
<end_of_turn>llama.cpp CLI
./llama-cli -m gemma-4-31B-Claude-4.6-Opus-Reasoning-Distilled-Q4_K_M.gguf -p "Prove that the square root of 2 is irrational." -n 1024
3-bit
4-bit
5-bit
6-bit
8-bit
Base model
google/gemma-4-31B-it