Thank you for your reply! The snippet I pasted was sort of amalgamated. I was testing "nvidia/nemotron-4-340b-instruct". The exact code and erorr message is:
from llama_index.llms.nvidia import NVIDIA
Settings.llm = NVIDIA(model="nvidia/nemotron-4-340b-instruct", ...)
INFO:httpx:HTTP Request: POST https://integrate.api.nvidia.com/v1/chat/completions "HTTP/1.1 400 Bad Request"
2024-10-30 13:47:25.979 Uncaught app exception
Traceback (most recent call last):
...
/home/0/miniconda3/envs/gpu_rag/lib/python3.10/site-packages/openai/_base_client.py", line 1058, in _request
raise self._make_status_error_from_response(err.response) from None
openai.BadRequestError: Error code: 400 - {'type': 'about:blank', 'status': 400, 'title': 'Bad Request', 'detail': 'Inference error'}
I'm using llama-index-llms-nvidia 0.1.4 , but I can't upgrade to the latest on my project due to the dependency conflict. I tested, though, with
pip install -U llama-index-llms-nvidia --no-deps
but it gave the same error.
I'm not sure the upgrade with dependencies will fix the error. Thank you again!!