Find answers from the community

Home
Members
bradcohn
b
bradcohn
Offline, last seen 2 months ago
Joined September 25, 2024
I've been using the GPTWeaviateIndex previously but after my Weaviate database was wiped I've been unable to reupload everything. Curious if this is a bug or I'm just trying to upload too large of an index at one time. Here's the error I'm getting.

Plain Text
File ~/projects/GPeaT/backend/venv/lib/python3.8/site-packages/requests/sessions.py:701, in Session.send(self, request, **kwargs)
    698 start = preferred_clock()
    700 # Send the request
--> 701 r = adapter.send(request, **kwargs)
    703 # Total elapsed time of the request (approximately)
    704 elapsed = preferred_clock() - start

File ~/projects/GPeaT/backend/venv/lib/python3.8/site-packages/requests/adapters.py:578, in HTTPAdapter.send(self, request, stream, timeout, verify, cert, proxies)
    576     raise SSLError(e, request=request)
    577 elif isinstance(e, ReadTimeoutError):
--> 578     raise ReadTimeout(e, request=request)
    579 elif isinstance(e, _InvalidHeader):
    580     raise InvalidHeader(e, request=request)

ReadTimeout: HTTPSConnectionPool(host='hv6yv9irrmanj7ha36vffa.gcp-b.weaviate.cloud', port=443): Read timed out. (read timeout=60)
5 comments
R
b
Hey -- I'm enjoying using llama-index very much for my project, and have it successfully deployed backing a chatbot. Now, I'm hoping to experiment and improve the model's responses to various questions, and if possible reduce the cost, as it's about 7c a query at the moment, which is too expensive for me to run publicly. Does anyone have experience with iterating through different parameterizations to improve model performance or reduce cost? I'm working with a substantial custom corpus (25MB) of somewhat high complexity.
4 comments
b
a
j
b
bradcohn
·

Deploying

I've got a nice chatbot/server working locally but I've been struggling to deploy it for a few days. It seems like the main issue is my index.json file is very large -- 175MB -- and so it's not compatible with the serverless/lambda pattern generally used by most of the platforms I'd like to deploy on. Does anyone have experience with this or advice on how to deploy apps that use llama_index?
15 comments
L
b
I've got a nice chatbot/server working locally but I've been struggling to deploy it for a few days. It seems like the main issue is my index.json file is very large -- 175MB -- and so it's not compatible with the serverless/lambda pattern generally used by most of the platforms I'd like to deploy on. Does anyone have experience with this or advice on how to deploy apps that use llama_index?
4 comments
b
j