Find answers from the community

Updated 4 months ago

Transitioning from LLM OpenAI to AzureOpenAI gpt4o deployment with token output limitations

At a glance

The community member transitioned from using OpenAI's LLM to Azure's OpenAI GPT-4 deployment, but they are unable to get the model to produce more than 1000 tokens. They have not set the max_tokens parameter and confirmed it is None in their settings. Other community members suggest that Azure may handle the lack of max_tokens setting differently than OpenAI, and recommend trying to set it to a value like 2000. However, this did not resolve the issue. One community member noted that the multimodal Azure LLM has a max_new_tokens setting of 300, which may be the cause of the limitation.

I transitioned from LLM OpenAI to AzureOpenAI gpt4o deployment, but I can't get the model to produce more than 1000 tokens. I have not set up max_tokens and confirmed its None in Settings.llm. Not sure what settings im missing here. any one experience the same?
O
L
5 comments
Pretty sure if max tokens isn't set, it sends it as None to the api. Maybe openai handles that differently than azure? Have you tried actually setting it to a value like 2000?
yeah, doesnt make a difference
i think it was the multimodal azure llm, it has max_new_tokens set to 300
πŸ‘€ ohhh multimodal
Add a reply
Sign up and join the conversation on Discord