embeddings change the "retrieval" step -- essentially, embeddings are what is used to compare a query to all the data you indexed, and retrieves the top-k most similar
Then, using the retrieved text, the LLM handles synthesizing a response to the query in natural language