Find answers from the community

Home
Members
kush2861
k
kush2861
Offline, last seen 3 months ago
Joined September 25, 2024
k
kush2861
·

Index update

Hello. Can I update an existing index with new data without generating embeddings for the old data again?
7 comments
W
k
Z
k
kush2861
·

CSV

Are there are some good solutions to RAG over CSV?
1 comment
W
The simple VectorStoreIndex.from_documents() uses what architecture? HNSW, ANN, IVF etc?
1 comment
L
Does anyone else have a probkem running perplexity with llama_index? I always get http 400 error?
7 comments
k
W
Is there any help for implementing RAG papers in Llama-Index? I want to reproduce the HippoRAG paper.
https://github.com/OSU-NLP-Group/HippoRAG
11 comments
b
L
k
Is there any implementation of DPR and DPRV2 retrievers in LlamaIndex?
1 comment
L
k
kush2861
·

Pdf

Does the pdf reader ignore the images in the file?
2 comments
k
L
k
kush2861
·

Embeddings

Are the new openai embedding models supported in llamaindex query engine?
Plain Text
ValueError: shapes (1536,) and (3072,) not aligned: 1536 (dim 0) != 3072 (dim 0)
15 comments
k
L
k
kush2861
·

Entity

Plain Text
entity_extractor = EntityExtractor(prediction_threshold=0.2,label_entities=False, device="cpu")

        node_parser = SentenceSplitter(chunk_overlap=200,chunk_size=2000)

        transformations = [node_parser, entity_extractor]

        documents = SimpleDirectoryReader(input_dir=r"Text_Files").load_data()

        pipeline = IngestionPipeline(transformations=transformations)

        nodes = pipeline.run(documents=documents)

        service_context = ServiceContext.from_defaults(llm=OpenAI(model="gpt-3.5-turbo", temperature=0),embed_model=embed_model)

        index = VectorStoreIndex(nodes, service_context=service_context)
Can I speed up this entity extraction process? It's very slow. Takes about an hour or so for 300 files.
3 comments
L
k
When I pass an embedding model to the service context and that service context to query engine, I am just changing embeddings of the query, right? I am loading index form storage.
6 comments
k
L
R
Is there any way to pass retrieved nodes to the query engine instead of the retriever. Asking this because I want to process the retrieved nodes before querying.
2 comments
k
L
I am using two different approaches in advanced RAG. I want to switch to option 2 in case 1 fails. I am depending on the response string and using an if else construct to switch between the approaches. For example if "I'm sorry but..." in response.response: use_approach_2. Is there any other way to do this?
9 comments
k
W
k
kush2861
·

Nodes

I am using a query engine. I want to query in such a manner that once a query is answered, the nodes used to answer that query should be dropped from the index before answering the next query and so on. Is it possible?
10 comments
k
W
I am using entity extractor for my data. I created the vector index and persisted locally. now while querying I am not loading the index from persists directory and not creating the nodes again. Can I find out what entities are retrieved for a particular query while doing so?
3 comments
n
k
is there distances_from_embeddings calculator in llamaindex like openai. the openai one is removed.
7 comments
k
a
how can I use CondensePlusContextChatEngine with vision models? I want to pass images as well.
6 comments
L
k
Which gpt 4 version does the default llama index setting point to?
1 comment
L
Is workflows analogous to lang graph in Langchain?
2 comments
L
k
kush2861
·

Splade

Is there a SPLADE implementation in Llama Index?
3 comments
k
L
k
kush2861
·

Embeddings

Why is it generating embeddings so many times?
1 comment
L
Is there any reason for chunk sizes to be powers of 2?
7 comments
k
L
L
Anyone getting this error while using BM25 Retriever?
Plain Text
ValidationError: 1 validation error for NodeWithScore node Can't instantiate abstract class BaseNode with abstract methods get_content, get_metadata_str, get_type, hash, set_content (type=type_error)
18 comments
k
L
W
I am trying to find where the query embeddings are created once I enter my query but I can't seem to find it. I thought it is using
Plain Text
get_top_k_embeddings
but when I print query_embeddings_np nothing appears in terminal. https://github.com/run-llama/llama_index/blob/main/llama-index-core/llama_index/core/indices/query/embedding_utils.py
3 comments
k
L
W
Is there someway by which I can use euclidean distance for similarity rather than cosine for retrieval using llamaindex?
2 comments
W
T
Is there any file in the llama index repo where I can see the implementation dot product when we use the vector retriever?
3 comments
k
L