Find answers from the community

Updated 4 months ago

How will I do that for llama2, will I

At a glance

The community member is asking how to use the LLama2 model, specifically whether they need to host the model themselves or if there are services that already host open-source models. The comments suggest that the community member can try using HuggingFace or Replicate, which are services that provide access to open-source language models like LLama2. The comments explain that the community member would need to get an API key to use these services, and provide example integrations for using LLama2 with Replicate and HuggingFace. The comments also describe Replicate as an inference service for large language models, which can be used to query open-source models instead of using a service like OpenAI.

Useful resources
How will I do that for llama2, will I have to first host my llama2 model somewhere, are there no services out there that already hosts open source models, so I can just use their API
a
B
8 comments
hey no silly question here:

you can try using huggingface or replicate for open source llms
you'd need to get an api key in order to use their services, but after you have that, you can use our LLM integrations for either of the two

What does replicate do exactly?
Is it a cloud compute service for LLMs?
Thanks for replying btw
Yup, they're an inference service for LLMs https://replicate.com/
they even offer fine-tuning services too
so instead of hitting openai's API for example to query an LLM, you can use replicate in similar ways to query an open-sourced LLM of your choosing (by providing a model name to any of the models they support)
Add a reply
Sign up and join the conversation on Discord