How will I do that for llama2, will I have to first host my llama2 model somewhere, are there no services out there that already hosts open source models, so I can just use their API
so instead of hitting openai's API for example to query an LLM, you can use replicate in similar ways to query an open-sourced LLM of your choosing (by providing a model name to any of the models they support)