- LlamaParse is interacting with LlamaCloud so nothing happens on your machine actually.
You give the file, it sends the file to llamacloud , gets processed and the final response is returned.
There is async way for multiple files cases, but the process remains the same.
- Yeah you can use custom LLM class and wrap it around your llm and use it.
- Are you using OpenAI and still getting resposne in 26 seconds?