I was using ollama to do the property graph construction, but it was too slow. I got an exl2 quant, installed exllama, and installed tabbyAPI. when I try to use llm = OpenAILike(model="text2cypher-codestral-exl2-4.0bpw", api_base="http://127.0.0.1:5000/v1/", api_key="fake", temperature=0.1, top_p=0.1, top_k=40, repetition_penalty=1.18), it only passes the temperature, nothing else.
also, it seems that the model runs forever when using this (if I say what is your name, it will start asking itself multiple questions after that; I assume this is a settings issue?? I did not have this problem with ollama).
when I try to run
kg_extractor = SchemaLLMPathExtractor(
llm=llm,
possible_entities=entities,
possible_relations=relations,
kg_validation_schema=validation_schema,
strict=True,
max_triplets_per_chunk=5,
num_workers=1,
)
index = PropertyGraphIndex.from_documents(
documents,
embed_model=embed_model,
kg_extractors=[kg_extractor],
property_graph_store=graph_store,
show_progress=True,
)
it just runs forever and no api calls are made.
anyone have any experience with this openailike endpoint?