Skip to content

Conversation

@ZX-ModelCloud
Copy link
Collaborator

@ZX-ModelCloud ZX-ModelCloud commented Jul 2, 2024

reslove ISSUE#91

@ZX-ModelCloud ZX-ModelCloud changed the title transformers integration [FEATURE] Add Transformers Integration Jul 2, 2024
@ZX-ModelCloud ZX-ModelCloud marked this pull request as ready for review July 3, 2024 10:29
@Qubitium Qubitium merged commit 76b7fc9 into ModelCloud:main Jul 3, 2024
DeJoker pushed a commit to DeJoker/GPTQModel that referenced this pull request Jul 19, 2024
* add transformers integration

* use gptqmodel

* Update hf_quantizer_gptq.py

* add monkey_patch_gptq_transformers()

* cleanup

* Fix issue: incorrect qlinear in transformers inegration quantization model

* add unit tests of Transformers integration

* monkey patch model.save_pretrained()

* cleanup

* cleanup

* rename monkey_patch_gptq_transformers() to monkey_patch_gptqmodel_into_transformers()

* select_quant_linear() remove "disable_exllama" param

---------

Co-authored-by: Qubitium-ModelCloud <[email protected]>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants