-
Notifications
You must be signed in to change notification settings - Fork 15
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
TypeError:make mixbits quant linear()got an unexpected keyword argument 'device' #112
Comments
Hi @bg51717 Thanks for reporting this. It's been fixed. |
However, there are still bugs present when saving the model after quantizing it with GPTQ.
and the error stack is
I'm interested in model quantizing and I believe the QLLM is a great project.Thanks for your work! |
Hi @bg51717 I have test it locally, it should work. Could you give another shot on it?Thanks |
I have tried the previous commands,and have a new bug.
I also have tried the solution from microsoft/onnxruntime#20252 I still get the same error. |
|
The command executed successfully, but it seems that the final result did not pass. So, has it failed?
Besides,when I try to ues the other command :
it raise a error:
but I found "OptForCausalLM" in true_sequential_layers_for_model rather than 'OPTForCausalLM'.
And I want to know how to write the function like |
If you want to support a new model, Please read the original AWQ paper for more detail. |
Hi,@wejoncy . Thanks to your great work.I'm studying model quantization from this project. I would like to know if this project is currently complete? This is because I noticed there are 'todo' placeholders in the code and some discrepancies between function definitions and their usage.How complete are GPTQ, AWQ, and HQQ? |
Yeah, It's almost done in quantization functionality. Some TODOs are for code-clean/refactor. |
when I using
python -m qllm --model=/root/models/baichuan-inc/Baichuan2-7B-Base --method=gptq --nsamples=64 --wbits=4 --groupsize=128 --save /root/models/baichuan-inc/Baichuan2-7B-Base_gptq_4b --export_onnx /root/models/baichuan-inc/Baichuan2-7B-Base_gptq_4b_onnx/
,it raise a error:
The text was updated successfully, but these errors were encountered: