Find answers from the community

Updated last year

Custom

Hi Team, I’m using the Custom llm class and want to send out a streaming response. Basically we have a custom llm that hosted and I have an API endpoint that I need to call (Using requests in Python) and return a generator response for streaming. It’s a flask endpoint. Do we have any examples/ documentation for this? Thanks.
Add a reply
Sign up and join the conversation on Discord