Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[WIP] EoRA #1206

Draft
wants to merge 109 commits into
base: main
Choose a base branch
from
Draft

[WIP] EoRA #1206

wants to merge 109 commits into from

Conversation

Qubitium
Copy link
Collaborator

@Qubitium Qubitium commented Feb 3, 2025

  • add adapter property to QuantizeConfig + EoRA lora adapter/config

Sample quantization config format with EoRA extension:

quant_config = QuantizeConfig(
   bits=4,
   sym=False,
   group_size=128,
   adapter=EoRA(rank=128),
}

@Qubitium Qubitium marked this pull request as draft February 3, 2025 14:41
@nbasyl
Copy link
Collaborator

nbasyl commented Feb 4, 2025

Hi @Qubitium, @hutm is the person who developed the ExLlama EoRA kernel and will assist with the inference and validation tasks once I complete the first milestone. Could you grant him the write access to this branch? Thanks!

@Qubitium
Copy link
Collaborator Author

Qubitium commented Feb 4, 2025

Hi @Qubitium, @hutm is the person who developed the ExLlama EoRA kernel and will assist with the inference and validation tasks once I complete the first milestone. Could you grant him the write access to this branch? Thanks!

Done. Write access invite sent to @hutm.

@nbasyl
Copy link
Collaborator

nbasyl commented Feb 4, 2025

Hi @Qubitium, I am having error from line 186 in config.py when trying to import QuantizeConfig, the error is as follow:
image
Do you know how to resolve this?

I am using python3.10 and have installed all the required libraries following the readme

@Qubitium
Copy link
Collaborator Author

Qubitium commented Feb 4, 2025

Hi @Qubitium, I am having error from line 186 in config.py when trying to import QuantizeConfig, the error is as follow: image Do you know how to resolve this?

I am using python3.10 and have installed all the required libraries following the readme

@nbasyl Let me get slack installed so we can converse on slack. I should be able to fix this. This is strange error, almost as if our Type hint code is wrong. Not sure if we have python 3.10 compat bug.

ZX-ModelCloud and others added 30 commits February 12, 2025 14:59
# Conflicts:
#	gptqmodel/models/auto.py
#	gptqmodel/models/base.py
#	gptqmodel/nn_modules/qlinear/__init__.py
#	gptqmodel/nn_modules/qlinear/bitblas.py
#	gptqmodel/nn_modules/qlinear/dynamic_cuda.py
#	gptqmodel/nn_modules/qlinear/exllama.py
#	gptqmodel/nn_modules/qlinear/exllamav2.py
#	gptqmodel/nn_modules/qlinear/ipex.py
#	gptqmodel/nn_modules/qlinear/marlin.py
#	gptqmodel/nn_modules/qlinear/torch.py
#	gptqmodel/quantization/gptq.py
#	gptqmodel/utils/model.py
#	tests/test_dynamic.py
#	tests/test_eval.py
#	tests/test_perplexity.py
Signed-off-by: ZX-ModelCloud <zx@modelcloud.ai>
Signed-off-by: ZX-ModelCloud <zx@modelcloud.ai>
Signed-off-by: ZX-ModelCloud <zx@modelcloud.ai>
Signed-off-by: ZX-ModelCloud <zx@modelcloud.ai>
Signed-off-by: ZX-ModelCloud <zx@modelcloud.ai>
Signed-off-by: ZX-ModelCloud <zx@modelcloud.ai>
Signed-off-by: ZX-ModelCloud <zx@modelcloud.ai>
Signed-off-by: ZX-ModelCloud <zx@modelcloud.ai>
Signed-off-by: ZX-ModelCloud <zx@modelcloud.ai>
Signed-off-by: ZX-ModelCloud <zx@modelcloud.ai>
Signed-off-by: ZX-ModelCloud <zx@modelcloud.ai>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

5 participants