I was using ollama to do the property graph construction, but it was too slow. I got an exl2 quant, installed exllama, and installed tabbyAPI. when I try to use llm = OpenAILike(model="text2cypher-codestral-exl2-4.0bpw", api_base="http://127.0.0.1:5000/v1/", api_key="fake", temperature=0.1, top_p=0.1, top_k=40, repetition_penalty=1.18), it only passes the temperature, nothing else. also, it seems that the model runs forever when using this (if I say what is your name, it will start asking itself multiple questions after that; I assume this is a settings issue?? I did not have this problem with ollama).
index = PropertyGraphIndex.from_documents( documents, embed_model=embed_model, kg_extractors=[kg_extractor], property_graph_store=graph_store, show_progress=True, ) it just runs forever and no api calls are made.
anyone have any experience with this openailike endpoint?