OpenAI models does not support multi modal?
I tried the multimodal code example with gpt-40 and the output has no relation to the provided image while the code does not raise any error.
Hi @MiladInk, yes multimodal inputs are currently disabled, but we're in the process of revamping our multimodal support (including images) in the near future! We'll update this issue when we have images (and other input modalities) working on OpenAI and other providers again :).
@nking-1
Hi @Harsha-Nori . Please, can you tell if this is effectively abandoned, and if yes, what was it abandoned in favor of? Is there another library that can do it better?
I've implemented a workaround that enables image support when using vLLM/OpenAI inference - https://github.com/guidance-ai/guidance/issues/1077