We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
b4273 for Ubuntu
Linux
CPU, CUDA
EPYC 7702, 3090
DeekSeek 2.5 1210
https://huggingface.co/deepseek-ai/DeepSeek-V2.5-1210
When running imatrix, I receive the error:
KV cache shifting is not supported for this model (--no-context-shift to disable)
However this error doesn't go away even if I pass that parameter:
./llama-imatrix -m /models_out/DeepSeek-V2.5-1210-GGUF/DeepSeek-V2.5-1210-Q8_0.gguf -f /training_dir/calibration_datav3.txt --output-file /models_out/DeepSeek-V2.5-1210-GGUF/DeepSeek-V2.5-1210.imatrix -t 120 --no-context-shift
Also tried with flash attention -fa to no avail
-fa
No response
llama_new_context_with_model: n_ctx_per_seq (512) < n_ctx_train (163840) -- the full capacity of the model will not be utilized llama_kv_cache_init: CPU KV buffer size = 2400.00 MiB llama_new_context_with_model: KV self size = 2400.00 MiB, K (f16): 1440.00 MiB, V (f16): 960.00 MiB llama_new_context_with_model: CPU output buffer size = 0.39 MiB llama_new_context_with_model: CUDA0 compute buffer size = 1422.00 MiB llama_new_context_with_model: CUDA_Host compute buffer size = 81.01 MiB llama_new_context_with_model: graph nodes = 4480 llama_new_context_with_model: graph splits = 1080 (with bs=512), 1 (with bs=1) common_init_from_params: KV cache shifting is not supported for this model (--no-context-shift to disable)' main : failed to init
The text was updated successfully, but these errors were encountered:
Adding LLAMA_EXAMPLE_IMATRIX to --no-context-shift resolves the issue, I assume that's a reasonable solve?
LLAMA_EXAMPLE_IMATRIX
--no-context-shift
Sorry, something went wrong.
Successfully merging a pull request may close this issue.
Name and Version
b4273 for Ubuntu
Operating systems
Linux
GGML backends
CPU, CUDA
Hardware
EPYC 7702, 3090
Models
DeekSeek 2.5 1210
https://huggingface.co/deepseek-ai/DeepSeek-V2.5-1210
Problem description & steps to reproduce
When running imatrix, I receive the error:
KV cache shifting is not supported for this model (--no-context-shift to disable)
However this error doesn't go away even if I pass that parameter:
./llama-imatrix -m /models_out/DeepSeek-V2.5-1210-GGUF/DeepSeek-V2.5-1210-Q8_0.gguf -f /training_dir/calibration_datav3.txt --output-file /models_out/DeepSeek-V2.5-1210-GGUF/DeepSeek-V2.5-1210.imatrix -t 120 --no-context-shift
Also tried with flash attention
-fa
to no availFirst Bad Commit
No response
Relevant log output
The text was updated successfully, but these errors were encountered: