Log in
Log into community
Find answers from the community
View all posts
Related posts
Did this answer your question?
๐
๐
๐
Powered by
Hall
Inactive
Updated 3 months ago
0
Follow
Ok next issue ๐ trying to get streaming
Ok next issue ๐ trying to get streaming
Inactive
0
Follow
b
big_ol_tender
last year
ยท
Ok next issue ๐ trying to get streaming to work. Im using a langchain llm (HuggingFaceTextGenInference) and the streaming works from my inference endpoint. However, when using it with llama_index I get error โLLM must support streamingโ
L
b
8 comments
Share
Open in Discord
L
Logan M
last year
wow thats annoying lol
Our streaming looks for an attribute called "streaming", but that particular LLM uses an attribute called "stream" ๐คฆโโ๏ธ
L
Logan M
last year
One quick workaround is this
Plain Text
Copy
class LlamaHFTextGen(HuggingFaceTextGenInference): streaming: bool = True
L
Logan M
last year
Basically override the langchain class to add that parameter
L
Logan M
last year
75% sure that will work lol
b
big_ol_tender
last year
Thanks! I actually just got it to work- I set streaming = True in the huggingfacetextgeninference class and NOWHERE else!
L
Logan M
last year
well, that works too! lol
b
big_ol_tender
last year
Yeah, and it streams for both retrieverqueryengine and index.as_chat_engine so far
b
big_ol_tender
last year
About to try others
Add a reply
Sign up and join the conversation on Discord
Join on Discord