This repository has been archived by the owner on Jun 24, 2024. It is now read-only.
Update to latest upstream BLOOM implementation / BLOOM Quantization does not work #228
Labels
Milestone
I am able to perform inference using the f16 model found here https://huggingface.co/nouamanetazi/bloomz-560m-ggml/tree/main but when I use
llm
to quantize it to q4_0 format, the model produces gibberish.The text was updated successfully, but these errors were encountered: