| { | |
| "quantization": { | |
| "exclude_modules": [ | |
| "lm_head", | |
| "model.embed_tokens" | |
| ], | |
| "kv_cache_quant_algo": null, | |
| "quant_algo": "FP8" | |
| } | |
| } | 
| { | |
| "quantization": { | |
| "exclude_modules": [ | |
| "lm_head", | |
| "model.embed_tokens" | |
| ], | |
| "kv_cache_quant_algo": null, | |
| "quant_algo": "FP8" | |
| } | |
| } |