make fp8 model quantized by llm-compressor can be inferenced in turbomind#4509
Open
43758726 wants to merge 3 commits intoInternLM:mainfrom
Open
make fp8 model quantized by llm-compressor can be inferenced in turbomind#450943758726 wants to merge 3 commits intoInternLM:mainfrom
43758726 wants to merge 3 commits intoInternLM:mainfrom