Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add GPTQModel support for evaluating GPTQ models #2217

Open
wants to merge 21 commits into
base: main
Choose a base branch
from

Conversation

Qubitium
Copy link

@Qubitium Qubitium commented Aug 16, 2024

Add option to use GPTQModel for lm_eval. GPTQModel is a replacement for AutoGPTQ on GPTQ quantization and inference with better model support and much faster inference speed out-of-the-box. We have been using it internally with lm-eval for months without issue.

@CLAassistant
Copy link

CLAassistant commented Aug 16, 2024

CLA assistant check
All committers have signed the CLA.

@Qubitium
Copy link
Author

@baberabb Hi, can we get some action on this? What do we need to do to get this reviewed and merged?

@Qubitium
Copy link
Author

@baberabb Ruff/Lint checks passed. Awaiting review. Thanks.

@Qubitium
Copy link
Author

@baberabb Ping. Please check our PR. We will push a unit test into test/modes/test_gptq.py later today to complete the PR. Let us know if there is anything else required of us.

@Qubitium Qubitium changed the title Add GPTQModel support for inferencing GPTQ models Add GPTQModel support for evaluating GPTQ models Oct 24, 2024
@Qubitium
Copy link
Author

@baberabb Unit test added.

@baberabb
Copy link
Contributor

Hi! Thanks for the PR, and sorry it took ages for us to review. This looks good to me, but I want to run it through @haileyschoelkopf as well.

ps. 3.8 test failing because of a recent transformers update.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

4 participants