llama_index.set_global_handler("simple")
I am not seeing any verbosity/debug messages for building the vector store indexTraceback (most recent call last): File "/opt/app-root/src/llamaindex-rag-example/starter.py", line 8, in <module> embed_model = HuggingFaceEmbedding(model_name="Cohere/Cohere-embed-english-v3.0") ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/opt/app-root/lib64/python3.11/site-packages/llama_index/embeddings/huggingface.py", line 82, in __init__ model = AutoModel.from_pretrained( ^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/opt/app-root/lib64/python3.11/site-packages/transformers/models/auto/auto_factory.py", line 526, in from_pretrained config, kwargs = AutoConfig.from_pretrained( ^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/opt/app-root/lib64/python3.11/site-packages/transformers/models/auto/configuration_auto.py", line 1132, in from_pretrained raise ValueError( ValueError: Unrecognized model in Cohere/Cohere-embed-english-v3.0. Should have a `model_type` key in its config.json, or contain one of the following strings in its name: ...
BAAI/bge-base-en-v1.5
) relate? or does it maybe not relate?ValueError: shapes (1024,) and (768,) not aligned: 1024 (dim 0) != 768 (dim 0)
Retrying llama_index.embeddings.openai.base.get_embeddings in 0.275948059787665 seconds as it raised AuthenticationError: Error code: 401 - {'error': {'message': 'Incorrect API key provided: xxx. You can find your API key at https://platform.openai.com/account/api-keys.', 'type': 'invalid_request_error', 'param': None, 'code': 'invalid_api_key'}}.
** Prompt **
and ** Completion **
stuffContext information is below. --------------------- file_name: summary-docs/cluster-autoscaling.md
UnboundLocalError: local variable 'default_template' referenced before assignment
print(summary.response)
and it looks like it's doing a double-summarySummaryIndex
via a TGIS server (and not run the LLM locally) but llamaindex seems like it's ignoring the TGIS predictor. Maybe I'm using this wrong?service_context = ServiceContext.from_defaults(chunk_size=512, llm=tgis_predictor, context_window=2048, prompt_helper=prompt_helper, embed_model=embed_model) # Load data documents = SimpleDirectoryReader('private-data').load_data() index = SummaryIndex.from_documents(documents) summary = index.as_query_engine(response_mode="tree_summarize").query("Summarize the text, describing what it might be most useful for")
Downloading url https://huggingface.co/TheBloke/Llama-2-13B-chat-GGML/resolve/main/llama-2-13b-chat.ggmlv3.q4_0.bin to path /tmp/llama_index/models/llama-2-13b-chat.ggmlv3.q4_0.bin total size (MB): 7323.31
import torch
or set the torch
kwargs, does it default to using CPU, or will it automatically use GPU regardless?