-
Notifications
You must be signed in to change notification settings - Fork 16.2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
MM-ReAct (Multimodal Reasoning and Action) #2262
Comments
Is it available on langchain now? |
GPT-4(Vision) is now available which unlocks multi-modal reasoning and action. According to OpenAI:
This model is currently available using the web UI but API access is inevitable. |
llava (multimodal llama) is open sourced now |
It's true that LLaVA merged w/ But, we also need multi-modal integration w/ I put a ticket up for it: Worth keeping this open until that is in place. I merged a notebook that runs LLaVA ( |
Now, we can use |
According to OpenAI:
Since multimodal LLMs like GPT-4V are now generally available, I'm closing this issue |
For GPT-4, image inputs are still in limited alpha.
For GPT-3.5, it would be great to see LangChain use the MM-ReAct agent.
The text was updated successfully, but these errors were encountered: