We are trying to implement rag with Metadata Extractors and trying to implement mistral or deepinfra API. Is that possible to implement the same. please advise
this is trace i am getting llama index vllm implementation
Traceback (most recent call last): File "/notebooks/notebooks/batch_proces/src/venv/lib/python3.9/site-packages/ray/_private/utils.py", line 527, in _get_docker_cpus cpu_ids.append(int(num_or_range)) ValueError: invalid literal for int() with base 10: '\n' 2024-01-11 03:11:51,986 INFO worker.py:1724 -- Started a local Ray instance. Traceback (most recent call last): File "/notebooks/notebooks/batch_proces/src/llms.py", line 14, in <module> llm = Vllm( File "/notebooks/notebooks/batch_proces/src/venv/lib/python3.9/site-packages/llama_index/llms/vllm.py", line 153, in init self._client = VLLModel( File "/notebooks/notebooks/batch_proces/src/venv/lib/python3.9/site-packages/vllm/entrypoints/llm.py", line 105, in init self.llm_engine = LLMEngine.from_engine_args(engine_args) File "/notebooks/notebooks/batch_proces/src/venv/lib/python3.9/site-packages/vllm/engine/llm_engine.py", line 307, in from_engine_args placement_group = initialize_cluster(parallel_config) File "/notebooks/notebooks/batch_proces/src/venv/lib/python3.9/site-packages/vllm/engine/ray_utils.py", line 112, in initialize_cluster raise ValueError( ValueError: The number of required GPUs exceeds the total number of available GPUs in the cluster.