Are there any good examples of using llama_index with a model on the Hugging Face Inference API? I know I'll load the model using llm = HuggingFaceHub(...), but (a) I seem to still need a local embedding model? and (b) Even when I use a local embedding model, I get "Empty Response" in an app where using llm = GPT4All(...) works well.