{ "quantization": { "method": "jang-importance", "profile": "JANG_4M", "target_bits": 4.0, "actual_bits": 5.1, "block_size": 64, "calibration_method": "weights", "quantization_method": "mse", "scoring_method": "weight-magnitude", "bit_widths_used": [ 4, 8 ], "quantization_scheme": "asymmetric", "quantization_backend": "mx.quantize" }, "source_model": { "name": "Gemma-4-31B-it-BF16", "dtype": "bfloat16", "parameters": "29.2B" }, "architecture": { "type": "transformer", "attention": "gqa", "has_vision": true, "has_ssm": false, "has_moe": false }, "runtime": { "total_weight_bytes": 19586875392, "total_weight_gb": 18.24 }, "format": "jang", "format_version": "2.0", "crack_surgery": { "method": "per-layer", "mode": "mpoa", "vector": "gemma4_31b_refusal_vectors.safetensors", "target_layers": [ 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49 ], "target_projs": [ "o_proj", "down_proj" ], "strength": 1.2, "modified_tensors": 60 } }