I'm using the multi_modal_pydantic notebook for pydantic + gpt-4 vision extraction and trying to adopt it for my use case.
Shouldn't the code below allow for multiple images to be loaded in so any images in restaurant_images can be called in sequence?:
from llama_index.multi_modal_llms import OpenAIMultiModal
from llama_index import SimpleDirectoryReader
# put your local directory here
image_documents = SimpleDirectoryReader("./restaurant_images").load_data()
openai_mm_llm = OpenAIMultiModal(
model="gpt-4-vision-preview", api_key=OPENAI_API_TOKEN, max_new_tokens=1000