{ "quant_method": "awq", "bits": 4, "group_size": 128, "dataset": "lmms-lab/flickr30k", "calibration_samples": 256, "model_seqlen": 2048, "block_name_to_quantize": "LlamaDecoderLayer", "module_name_preceding_first_block": [ "model.vision_model", "model.connector" ], "quantization_tool": "llm-compressor", "preserved_components": [ "vision_model", "vision_tower", "connector", "lm_head" ] }