Fixed quantization_config.llm_int8_skip_modules, to avoid re-quantizes embed_tokens layers on load 159a450 verified techwithsergiu commited on Mar 11