I'm guessing this is because the llm is a chat model, and it is converting my prompt input into a user chat message. Is this actually calling the LLM twice or is it just doing that prompt_to_messages conversion? And if it is calling twice, how can I just get the llm to call complete instead of chat when run through the QueryPipeline?
When trying to import AzureOpenAIEmbeddings I am getting an issue cannot import name root_validator from llama_index.core.bridge.pydantic. This appears to happen in the source code of the class itself? The code snippet in question does say that it is not used, any advice?