Find answers from the community

Updated 11 months ago

```

Plain Text



is this happen when you run out of memory?

im running dolphin-phi on 5600g amd apu using ollama
L
m
21 comments
I think it's just an http request timeout
You can increase the default request_timeout for ollama in the constructor
default is 30s
this is ollama or llama index limit?
Basically need to add llm = Ollama(..., request_timeout=60) or some other timeout value
llm = Ollama(model=self._model, base_url=self._base_url, request_timeout=300)

strange changed the base.py in the OllamaQueryEnginePack

stll time out in 30s

when rerunning

ollama_pack = OllamaQueryEnginePack(model="dolphin-phi", documents=documents)
response = ollama_pack.run("What is the title of the doc?")
What version of llama-index do you have?
v0.9.22 added the request_timeout kwarg
Plain Text
WARNING: Package(s) not found: langchain
Name: llama-index
Version: 0.9.25.post1
Name: transformers
Version: 4.36.2
Name: sentence-transformers
Version: 2.2.2
Name: pypdf
Version: 3.17.4
Note: you may need to restart the kernel to use updated packages.


python 3.11.5
Attachment
image.png
I'm at a loss for this one. As you can see in the traceback, the timeout is being passed in

Plain Text
    101 with httpx.Client(timeout=Timeout(self.request_timeout)) as client:
--> 102     response = client.post(
    103         url=f"{self.base_url}/api/chat",
    104         json=payload,
    105     )


We are even following httpx docs
https://www.python-httpx.org/advanced/#setting-a-default-timeout-on-a-client
This seems to require some rigorous debugging :PSadge:
Hmmm interesting 🧐
in the base.py can we comment out
llm = Ollama(model=self._model, base_url=self._base_url)

instead add

llm = MistralAI(api_key=api_key)

im moving to Mistral api to still built the rag while still debuging the timeout
if not do you know of any working rag pdf for any mistral cloud api demo. i need to get back on track and llocal llm with ollama is bugging out for the last week.
it should be as simple as swapping out the llm πŸ‘€
Plain Text



so ollama serve runs on boot as it was installed from their official script

when running the dolphin-phi its 8GB/16gb


it was working pretty good on m1 mac mini 8gb. lmfao fml
Attachment
image.png
@Logan M lmk if you see anything out of order.
whats the issue? πŸ‘€
(I feel blind lol but that looks fine to me?)
Add a reply
Sign up and join the conversation on Discord