{ "_class_name": "FluxTransformer2DModel", "_diffusers_version": "0.32.0", "model_type": "flux", "quantization_config": { "quant_method": "fp8", "quant_type": "e4m3fn" }, "base_model": "black-forest-labs/FLUX.2-klein-base-9B", "model_name": "BigLoveKleinFp8" }