@Logan M im currently trying to switch from llama.cpp to Ollama but the same model give me different responses. The output from Llama.cpp is correct and in the right language. The output from Ollama is wrong and sometimes in the wrong language... I have also talked to the Ollama community but we have no solution so far.... maybe it has to do with the implementation in llama index?
I have already compared all the settings i could find. I can provide you with whatever infos you need.
We could (from my viewpoint) greatly increase the quality of ollama if we could find out what is different.