-
-
Notifications
You must be signed in to change notification settings - Fork 14.7k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
ollama: 0.0.17 -> 0.1.7 #257760
ollama: 0.0.17 -> 0.1.7 #257760
Conversation
We need to allow to be built with gpu support before merge |
Note 0.1.1 is already available. |
Thanks, I've updated to 0.1.1 and also added CUDA/ROCm support (based on https://github.com/ggerganov/llama.cpp/blob/master/flake.nix) by integrating llama-cpp as a package. I don't have hardware to test CUDA and ROCm locally, maybe someone could try that out? |
Thank you for the effort, is there a reason that there is no opencl option (I think it should be supported and might be easier to setup on some devices than rocm and cuda), see Additionally I tested the the compiled binary with and without rocmSupport
With the version in nixpkgs/master it works (though without the option for gpu support) |
Rebased & updated ollama to 0.1.3. |
@benneti Thanks for testing. I'll look into that issue reg. ROCm. I've also added OpenCL support, could you test that as well? |
I've patched the passing of the deprecated --gqa flag, which seems to be only needed for GGML format models. Could you test that again? |
As I am unable (I run out of space while compiling rocblas and am too lazy to change my tmpfs setup for this) even with the fixes necessary (see code comment), I only tried the standard version and opencl support. |
I can also confirm that it now works as intended, |
The issue was fixed upstream, so I updated & removed the patch. |
I read through the change logs and it seems there are only additions and no major changes, and i don't consider this a breaking change for the 23.11 release. but the aarch64 darwin ofborg builder isn't doing too good, i think it could use some friends |
Version 0.0.17 is broken with the current models. This update puts it back in a working state. |
I can confirm that, actively using it on aarch64-darwin w/ Metal support. |
I installed ollama from unstable and I am getting:
Related upstream issues:
I am not sure, what's the issue. I have 11.8 toolkit running. Building ollama directly also doesn't seem to pickup the includes. I think this (ollama/ollama#958) should finally fix it? |
The override looks like this:
But it seems like compiling llma-cpp fails with issues mentioned here: ggerganov/llama.cpp#2481 EDIT: following worked ollamagpu = pkgs.unstable.ollama.override { llama-cpp = (pkgs.unstable.llama-cpp.override {cudaSupport = true; openblasSupport = false; }); }; |
Should we provide ollameWithNvidia and ollamaWithRocm in all packages like torch does? |
I would avoid to create |
Just wanted to thank @geekodour for his fix #257760 (comment) . Worked for me as well. |
Does this still work for you guys? I'm getting the following error:
And removing the disable cublas option then complains about running an unsupported compiler version. |
It stopped working for me so I removed the check altogether. Not sure what's the proper fix. |
So you're just using the CPU right? I removed the override and it installs just fine but holy hell is it slow lol. Must be some new configuration for llama-cpp I'm going to lookup if I have time later. |
I don't have a proper fix but ran into the following errors:
I don't think that first overlay is doing much of anything. Globally, cuda support is not turned on. Just thought I'd leave this here in case someone else stumbles on this issue. I'm also not using unstable. |
to try to address this, I made #277709 |
there has be a rewrite of the upstream nix files. It included a stdenv fix for cuda versions. (edit: oops, actually wanted to reply in #277709 , but here works too) |
Thank you! SomeoneSerge nicely commented on what had to change, I think all is incorporated now. Let me know if you notice something I missed. |
Description of changes
Updated the package and added llama-cpp as a required dependency. Also added support for Metal, CUDA, ROCm hw acceleration.
Things done
sandbox = true
set innix.conf
? (See Nix manual)nix-shell -p nixpkgs-review --run "nixpkgs-review rev HEAD"
. Note: all changes have to be committed, also see nixpkgs-review usage./result/bin/
)