Skip to content

make fp8 model quantized by llm-compressor can be inferenced in turbomind#4509

Open
43758726 wants to merge 3 commits intoInternLM:mainfrom
43758726:add/llm-compressor-fp8-inference
Open

make fp8 model quantized by llm-compressor can be inferenced in turbomind#4509
43758726 wants to merge 3 commits intoInternLM:mainfrom
43758726:add/llm-compressor-fp8-inference

Commits

Commits on Apr 18, 2026