Find answers from the community

t
tarpus
Offline, last seen 3 weeks ago
Joined September 25, 2024
t
tarpus
·

Vision

What is the best vision model currently that performs as well as gpt4o? That is an LLM with vision modality.
7 comments
L
t
t
tarpus
·

Image

does LI have a solution to managing the size of a page to send the oai image model?
to manage resolution vs. tokens?
7 comments
L
t
InvalidRequestError: The model text-davinci-003 has been
4 comments
L
t
Does title extractor work with version .8.43
8 comments
L
t
I may be doing something wrong with respect to getting line numbers.

I am not getting the most accurate citations when it comes down to the line number. The page numbers seem spot on. Is there an art to fetching the line number that is significantly different than getting the page number?
8 comments
t
T
Does llama index support agent that can read data from and write data to a SQL db (postgres , sqlite, etc.)?
5 comments
L
t
I'm guessing that the they are now

from llama_index.readers.file import PyMuPDFReader
and this
from llama_index.core.node_parser import SentenceSplitter respectively
1 comment
L
what happened to this:

from llama_index.vector_stores import PineconeVectorStore vector_store = PineconeVectorStore(pinecone_index=pinecone_index)

ModuleNotFoundError: No module named 'openai.openai_object'
100 comments
L
t
W
any idea how I can extend this to PDFs instead of web pages?
7 comments
t
L
@Logan M bro are LI agents in any way easier to use than langchain 🙂
1 comment
W
I can't believe I'm asking this, but I have no idea now how to call the open ai api any longer
11 comments
W
t
T
does anyone know the solution to this very quickly. This no longer works/ deprecated

response = openai.ChatCompletion.create(
model='gpt-4-1106-preview',
messages=messages
)
3 comments
W
t
s
t
tarpus
·

Product

question - sometimes I see products featured on Llama Index thread on Linkedin. How can we showcase a prodcut?
2 comments
t
L
What do you guys think of how Gemini will change things?
5 comments
L
t
do you guys have any solution to sending a pdf to gpt vision api?
9 comments
t
L
I'm writing an article about my experience with Llama index, amongst other technologies. I am going to post it tomorrow, most likely - will have a link to your site 🙂

While not in the millions of followers - I do have 10's of thousands so let's see what happens. Any advice about how to call you guys out (in addition to a link to your site?) of which I should be mindful?
5 comments
t
L
I am looking for API docs for llama index in PDF form
2 comments
t
L
for the sec-insights UI, can I just use my own RAG code, and leverage the UI as just a front-end?
5 comments
b
L
t
I am able to get the knowledge graph page citations. but I can't get the document name to print:

line_number = 1
documents = []
for doc_idx, (chunk, page_number) in enumerate(chunks):
line_count_in_chunk = chunk.count('\n') + 1
metadata = {
"source_doc_idx": doc_idx,
"filename": "1.2.2.2 Customer Contract - Stockwood Dr - Woodstock - GA.pdf",
"page_number": page_number,
"document_title": "Customer Contract - Stockwood Dr - Woodstock - GA",
"line_count": line_count_in_chunk,
"starting_line_number": line_number
}
documents.append(Document(text=chunk, metadata=metadata))
line_number += line_count_in_chunk

for doc in documents:
print(f"Document Metadata: {doc.metadata}")

service_context = ServiceContext.from_defaults(
llm=OpenAI(model="gpt-3.5-turbo", temperature=0)
)

file_path = "./1.2.2.2 Customer Contract - Stockwood Dr - Woodstock - GA.pdf"
file_name = os.path.basename(file_path)

try:
index = VectorStoreIndex.from_documents(nodes, service_context=service_context)
except Exception as e:
print(f"Error: {e}")

query_engine = CitationQueryEngine.from_args(
index,
similarity_top_k=3,
citation_chunk_size=512,
)

response = query_engine.query("what is the purchase commitment?")
print("Query Response:", response)

G = nx.Graph()
for i, source_node in enumerate(response.source_nodes):
node_content = source_node.node.get_text()
citation_page = documents[i].metadata['page_number']
G.add_node((file_name, citation_page), content=node_content)
8 comments
t
L
t
tarpus
·

Metadata

I then leverage llama-index, and get bogus page number
20 comments
t
L
T