You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
The best workaround we have, if you specifically need to use the grounding feature, is to do two separate calls. One to get information about the image/document (without grounding), then another with the image description with grounding enabled.
Is your feature request related to a problem? Please describe.
I'm frustrated that VertexAI grounding is not supported if input is non-text.
Describe the solution you'd like
Is there currently a convenient workaround for it? I'd like to have this functionality of being able to ask for ex:
input(what this item is made of? + [image]) -> grounded search of what item like this is usually made of in the document -> output(text)
Describe alternatives you've considered
There is obviously a way of making 2 separate calls:
However, that increases costs massively.
Additional context
No response
Code of Conduct
The text was updated successfully, but these errors were encountered: