I would recommend you go through create-llama project from LlamaIndex.
It will help you deploy your full application just based on your the basic requirements like what llm do you want to use.
What vector store if you want to use.
What framework ( for backend + frontend )
You can opt for opensource llms which will run locally.
https://www.llamaindex.ai/blog/create-llama-a-command-line-tool-to-generate-llamaindex-apps-8f7683021191