Nemotron-Cascade-2-30B-A3B โ€” Q5_1 GGUF

GGUF quantization of nvidia/Nemotron-Cascade-2-30B-A3B.

  • Architecture: Hybrid Attention + Mamba (SSM) + MoE โ€” 30B total parameters, 3B active
  • Quantization: Q5_1 (uniform 5-bit with delta and min per block)

Quantization commands

# Convert HF model to GGUF (bf16)
python llama.cpp/convert_hf_to_gguf.py \
  nvidia/Nemotron-Cascade-2-30B-A3B \
  --outfile nemotron-cascade-30b-bf16.gguf \
  --outtype bf16

# Quantize to Q5_1
llama-quantize nemotron-cascade-30b-bf16.gguf \
  nemotron-cascade-30b-Q5_1.gguf Q5_1

Usage

Load in LM Studio, llama.cpp, or any GGUF-compatible runtime.

Downloads last month
186
GGUF
Model size
32B params
Architecture
nemotron_h_moe
Hardware compatibility
Log In to add your hardware

5-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for AdrienBrault/Nemotron-Cascade-2-30B-A3B-Q5_1-GGUF

Quantized
(31)
this model