-
-
Notifications
You must be signed in to change notification settings - Fork 381
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Inference Failed with Ollama #20
Comments
Ollama is a new addition (a couple of hours old), so it's not as battle tested as using openAI. I'll probably add I managed to reproduce some weird behaviors at least with the |
Added some warnings in 5cbce67. |
For sure these are not high priority issues. Just trying to add issues as I see them to help with your development. Liking what I'm seeing so far. I just got a new one that had more data in the response: |
Keep the bug reports coming, they are really helpful! Yes, that's the error I got as well. As I mentioned, the model sometimes decides to ignore my instruction and outputs either an invalid json or a random json :D This is what casues this error. Btw, I'm curious, which model are you using? |
Yes, using llama2. No other reason except that's what was in your documentation. I've used Ollama for about 2 hours at this point... :) |
I sent 8156e8c to solve the |
Not to jump onto your thread. But I've had two issues with this as well. First: Second: |
@dyslexicjedi yeah, that one is the model not respecting the prompt. Sometimes the models just forgets about the prompt (I guess with long contexts?) and starts just outputing random JSONs. I wonder if there are some techniques that I can use for the model not to lose context. For example, Maybe I share the content first, and then ask the model to give me the tags? 🤔 |
Changed the prompt to be split in 2, so ollama does not forget them
@dyslexicjedi Seems like I had the same idea as MohamedBassem. I tried it out before reading his answer and it seems to work fine. Basically I split the existing prompt into 2 and added the "respond with json and 3-5 tags" part at the end: For the samples above, the llama3 model responded properly:
I also tried it with the ~20 links I have in my hoarder instance so far and all of them were tagged (previously 3 weren't). |
@kamtschatka the change you shared in general looks good to me. I'm happy to accept it in a PR if this improves ollama's accuracy :) I can also give it a try a bit with openAI. |
Hi, I just set up Hoarder with OLLAMA. When it runs the inference jobs, i get errors such as: 2024-05-17T18:13:08.088Z info: [inference][68] Starting an inference job for bookmark with id "dzrjut9qb81couvfepd5pr7g" I don't know where else to look for deeper log data, so any input is greatly appreciated. I'm running the models: llama3 and llava-llama3 for image. Thanks! |
Sounds like it is not reachable, or running on the wrong port or something like that.
|
Thank you! That gave me enough direction to figure out what was going on. Getting successes now. |
* Inference Failed with Ollama #20 Changed the prompt to be split in 2, so ollama does not forget them * Update apps/workers/openaiWorker.ts Co-authored-by: Mohamed Bassem <me@mbassem.com> --------- Co-authored-by: kamtschatka <simon.schatka@gmx.at> Co-authored-by: Mohamed Bassem <me@mbassem.com>
As I continue to use the app with Ollama, I am having good success with the text based tagging using the llama3 stack. However, I am not having any of the image searches work, and they all create failed inferences. I've tried with llava and llava-llama3. I'm getting these errors in the workers log, but am at a loss to further troubleshoot. Any suggestions? 2024-05-22T16:23:33.588Z info: [inference][355] Starting an inference job for bookmark with id "z0t1rbmw8pslkc8rcjsp24zt" |
Anything I can do to help troubleshoot this? |
Provide instructions on what you are doing/have configured and a sample image. I just tried to upload an image and got proper tagging, so it can't be a general issue. |
It seems that ollama has been mostly stable with folks so far. So I'm closing this for now. |
I added ollama support per instructions. Some of the requests seemed to work fine but I found the following in the logs:
2024-03-27T14:01:40.649Z error: [inference][17] inference job failed: Error: Expected a completed response. 2024-03-27T14:06:40.997Z error: [inference][18] inference job failed: TypeError: fetch failed
At least one of my bookmarks does not have any tags added to it.
I'm running this on a docker container on linux.
The text was updated successfully, but these errors were encountered: