-
Notifications
You must be signed in to change notification settings - Fork 1.2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[BUG] helium vision_web_browser.py NoneType error after saving image #570
Comments
Same error with vision model : model = LiteLLMModel(
model_id="ollama_chat/llama3.2-vision:11b"
api_base="http://localhost:11434"
num_ctx=8192
) |
And in general it won't work unless this gets merged #553 But the only currently supported vision model llava is very bad. You can try using it with my PR: python src/smolagents/vision_web_browser.py --model-type LiteLLMModel --model-id=ollama/llava But it produces very bad results. In some cases the model does not return any python code on the first go and that kills execution. I created a PR which tries to fix that. |
Shouldn't this has been fixed by now @sysradium ? I tried again today with # Initialize the model based on the provided arguments
model = LiteLLMModel(
model_id="ollama_chat/llava:13b",# model_id="ollama_chat/mistral-small",
api_base="http://localhost:11434",
num_ctx=8192
) But I still encounter the following error :
|
@PrideIsLife yeah, as you can see they haven't merged it yet. But once it is merged it will still fail due to d34e0c8 which we are discussing here #655. But still you can overcome this disabling message flattening: model = LiteLLMModel(
model_id="ollama_chat/mistral-small",
api_base="http://localhost:11434",
num_ctx=8192
flatten_messages_as_text=False,
) |
Ok thank you @sysradium, BTW I just tested with model = LiteLLMModel(
# model_id="ollama_chat/llava:13b",# model_id="ollama_chat/mistral-small",
model_id="ollama_chat/mistral-small",
api_base="http://localhost:11434",
num_ctx=8192,
flatten_messages_as_text=False,
) and got the following error :
same with EDIT found some logs in ollama :
I guess disabling flattening increases vram usage ? |
@PrideIsLife the problem will go away only when my change gets merged :) |
To run vision_web_browser, you need to use a VLM model that supports images. You got this error message:
which means that your model does not support images. That is why, when your forced the image input format (
Please, try to use a VLM model that supports images. |
Describe the bug
When running vision_web_browser.py on Windows 11 with helium.
with the following model :
I get the error below.
Code to reproduce the error
src/smolagents/vision_web_browser.py
with the above model on Windows 11 with helium.Error logs (if any)
Expected behavior
No error
Packages version:
The text was updated successfully, but these errors were encountered: