-
Notifications
You must be signed in to change notification settings - Fork 252
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
awqint4 to gguf ,ModuleNotFoundError: No module named 'awq.apply_awq' #502
Comments
mport os model_path = 'mistralai/Mistral-7B-v0.1' model = AutoAWQForCausalLM.from_pretrained( model.quantize( model.save_quantized(quant_path) GGUF conversionprint('Converting model to GGUF...') if not os.path.exists(llama_cpp_path): subprocess.run([ subprocess.run([ |
Hi @LDLINGLINGLING. This seems to be a https://casper-hansen.github.io/AutoAWQ/examples/#gguf-export |
I didn't succeed,I followed the instructions in this link https://casper-hansen.github.io/AutoAWQ/examples/#gguf-export, but the error at the top appeared |
I now think this operation is meaningless, because I originally thought that awq has high quantization accuracy. Whether converting to gguf can maintain this accuracy, but it should be impossible |
Hi @LDLINGLINGLING ~ And by the way, I'm occur an error that might similar with this issue, hope someone can help me. I had already converted an Phi-3-mini-128K model to AWQ.
The error said that it cannot map through the define layer. If that so, how do I modify it? sorry for bad English, but hope someone can help. ;-; BR, Matt. |
I want to use awq quantize a model, and use llama.cpp convert to gguf. but I followed the tutorial but got an error:Traceback (most recent call last):
File "/root/ld/ld_project/llama.cpp/convert_minicpm.py", line 2516, in
main()
File "/root/ld/ld_project/llama.cpp/convert_minicpm.py", line 2460, in main
from awq.apply_awq import add_scale_weights # type: ignore[import-not-found]
ModuleNotFoundError: No module named 'awq.apply_awq'
my awq version is
autoawq 0.2.5+cu121
autoawq_kernels 0.0.6
The text was updated successfully, but these errors were encountered: