RAM optimised Gemma 4 models by baa.ai
AI & ML interests
Model Quantization
Recent Activity
View all activity
MINT & SWAN quantized versions of MiniMax-M2.5 (MLX & GGUF)
SWAN quantized versions of Llama 3.1 and 3.3 70B Instruct (MLX)
MINT & SWAN quantized versions of Qwen3 models (MLX)
MINT quantized versions of Qwen3.5-122B-A10B at multiple budget targets (MLX & GGUF)
MINT quantized Nemotron-3-Super-120B — hybrid Mamba-MoE-Attention (MLX & GGUF)
Baa.ai quantized versions of GLM models
MINT & SWAN quantized versions of Llama 4 Scout and Maverick (MLX & GGUF)
MINT quantized versions of Qwen3.5-35B-A3B at multiple budget targets (MLX & GGUF)
MINT & SWAN quantized versions of Qwen3.5-397B-A17B (MLX & GGUF)
RAM optimised Gemma 4 models by baa.ai
MINT quantized Nemotron-3-Super-120B — hybrid Mamba-MoE-Attention (MLX & GGUF)
MINT & SWAN quantized versions of MiniMax-M2.5 (MLX & GGUF)
Baa.ai quantized versions of GLM models
SWAN quantized versions of Llama 3.1 and 3.3 70B Instruct (MLX)
MINT & SWAN quantized versions of Llama 4 Scout and Maverick (MLX & GGUF)
MINT & SWAN quantized versions of Qwen3 models (MLX)
MINT quantized versions of Qwen3.5-35B-A3B at multiple budget targets (MLX & GGUF)
MINT quantized versions of Qwen3.5-122B-A10B at multiple budget targets (MLX & GGUF)
MINT & SWAN quantized versions of Qwen3.5-397B-A17B (MLX & GGUF)