Update GPT-QModel references and deprecate AutoGPTQ#3190
Update GPT-QModel references and deprecate AutoGPTQ#3190BenjaminBossan merged 8 commits intohuggingface:mainfrom
Conversation
|
@BenjaminBossan Conflict resolved and we have tested this PR and relevant tests are passing and does not rely on the Optimum PR to be merged first. This one is ready for merge. Please trigger ci tests to make sure there are no regressions. |
BenjaminBossan
left a comment
There was a problem hiding this comment.
Thanks for the PR to fully remove AutoGPTQ and move to GPT-QModel.
The PR generally looks good and both single and multi GPU tests passed in my testing. There are a few smaller issues left, please check.
| ```bash | ||
| # gptqmodel install | ||
| pip install gptqmodel --no-build-isolation | ||
| # GPTQ-Model install |
There was a problem hiding this comment.
| # GPTQ-Model install | |
| # GPT-QModel install |
|
@BenjaminBossan Ready for ci tests re-run.
|
| # a NVIDIA L4). So we fix the compute capability to 8.9. In the future we might extend this | ||
| # to a list of compute capabilities (separated by ;). | ||
| RUN CUDA_ARCH_LIST=8.9 conda run -n peft pip install --no-build-isolation "gptqmodel>=6.0.3" | ||
| RUN CUDA_ARCH_LIST=8.9 conda run -n peft pip install "gptqmodel>=7.0.0" |
There was a problem hiding this comment.
7.0.0 no longer requires --no-build-isolation as all kernels have moved from setup stage to first-use JIT.
|
The docs for this PR live here. All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update. |
BenjaminBossan
left a comment
There was a problem hiding this comment.
Thanks for the update, everything looks good. Failing CI is unrelated and I ran the GPU tests locally to ensure that they pass too.
And congrats to the 7.0 release.
@BenjaminBossan
GPT-QModelas the pkg is no longergptqonly, it now fully supports awq and many other quant methods.Companion (but not required) PR in Optimum: huggingface/optimum#2426